Search Results for author: Heinrich Peters

Found 5 papers, 0 papers with code

Social Media Use is Predictable from App Sequences: Using LSTM and Transformer Neural Networks to Model Habitual Behavior

no code implementations20 Apr 2024 Heinrich Peters, Joseph B. Bayer, Sandra C. Matz, Yikun Chi, Sumer S. Vaid, Gabriella M. Harari

Additionally, our analyses reveal that the person-level predictability of social media use is not substantially related to the frequency of smartphone use in general or the frequency of social media use, indicating that our approach captures an aspect of habits that is distinct from behavioral frequency.

Context-Aware Prediction of User Engagement on Online Social Platforms

no code implementations23 Oct 2023 Heinrich Peters, Yozen Liu, Francesco Barbieri, Raiyan A. Baten, Sandra C. Matz, Maarten W. Bos

The success of online social platforms hinges on their ability to predict and understand user behavior at scale.

Privacy Preserving

Generalizable Error Modeling for Search Relevance Data Annotation Tasks

no code implementations8 Oct 2023 Heinrich Peters, Alireza Hashemi, James Rae

This paper presents a predictive error model trained to detect potential errors in search relevance annotation tasks for three industry-scale ML applications (music streaming, video streaming, and mobile apps) and assesses its potential to enhance the quality and efficiency of the data annotation process.

Model Share AI: An Integrated Toolkit for Collaborative Machine Learning Model Development, Provenance Tracking, and Deployment in Python

no code implementations27 Sep 2023 Heinrich Peters, Michael Parrott

AIMS features collaborative project spaces and a standardized model evaluation process that ranks model submissions based on their performance on unseen evaluation data, enabling collaborative model development and crowd-sourcing.

Large Language Models Can Infer Psychological Dispositions of Social Media Users

no code implementations13 Sep 2023 Heinrich Peters, Sandra Matz

As Large Language Models (LLMs) demonstrate increasingly human-like abilities in various natural language processing (NLP) tasks that are bound to become integral to personalized technologies, understanding their capabilities and inherent biases is crucial.

Zero-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.