physics.soc-ph
42 postsarXiv:2503.09811v1 Announce Type: new Abstract: Understanding the mechanisms driving the distribution of scientific citations is a key challenge in assessing the scientific impact of authors. We investigate the influence of the preferential attachment rule (PAR) in this process by analyzing individual citation events from the DBLP dataset, enabling us to estimate the probability of citations being assigned preferentially. Our findings reveal that, for the aggregated dataset, PAR dominates the citation distribution process, with approximately 70% of citations adhering to this mechanism. However, analysis at the individual level shows significant variability, with some authors experiencing a greater prevalence of preferential citations, particularly in the context of external citations. In contrast, self-citations exhibit notably different behaviour, with only 20% following PAR. We also demonstrate that the prominence of PAR increases with an author's citability (average citations per paper), suggesting that more citable authors are preferentially cited, while less-cited authors experience more random citation patterns. Furthermore, we show that self-citations may influence bibliometric indexes. Our results emphasise the distinct dynamics of self-citations compared to external citations, raising questions about the mechanisms driving self-citation patterns. These findings provide new insights into citation behaviours and highlight the limitations of existing approaches in capturing the nuances of scientific impact.
arXiv:2408.07237v2 Announce Type: replace Abstract: Beliefs form the foundation of human cognition and decision-making, guiding our actions and social connections. A model encapsulating beliefs and their interrelationships is crucial for understanding their influence on our actions. However, research on belief interplay has often been limited to beliefs related to specific issues and relied heavily on surveys. We propose a method to study the nuanced interplay between thousands of beliefs by leveraging an online user debate data and mapping beliefs onto a neural embedding space constructed using a fine-tuned large language model (LLM). This belief space captures the interconnectedness and polarization of diverse beliefs across social issues. Our findings show that positions within this belief space predict new beliefs of individuals and estimate cognitive dissonance based on the distance between existing and new beliefs. This study demonstrates how LLMs, combined with collective online records of human beliefs, can offer insights into the fundamental principles that govern human decision-making.
arXiv:2410.04301v3 Announce Type: replace-cross Abstract: This work extends the recent opinion dynamics model from Cheng et al., emphasizing the role of group pressure in consensus formation. We generalize the findings to incorporate social influence algorithms with general time-varying, opinion-dependent weights and multidimensional opinions, beyond bounded confidence dynamics. We demonstrate that, with uniformly positive conformity levels, group pressure consistently drives consensus and provide a tighter estimate for the convergence rate. Unlike previous models, the common public opinion in our framework can assume arbitrary forms within the convex hull of current opinions, offering flexibility applicable to real-world scenarios such as opinion polls with random participant selection. This analysis provides deeper insights into how group pressure mechanisms foster consensus under diverse conditions.
arXiv:2404.02175v5 Announce Type: replace-cross Abstract: Understanding how consumers respond to business advertising efforts is essential for optimizing marketing investment. This research introduces a new modeling approach based on the concepts of symmetries and scaling laws in physics to describe consumer response to advertising dynamics. Drawing from mathematical frameworks used in physics and social sciences, we propose a model that accounts for a key aspect: the saturation effect. The model is validated against commonly used models, including the Michaelis-Menten and Hill equations, showing its ability to better capture nonlinearities in advertising effects. We introduce new key parameters like Marketing Sensitivity, Response Sensitivity, and Behavioral Sensitivit, that offer additional insights into the drivers of audience engagement and advertising performance. Our model provides a rigorous yet practical tool for understanding audience behavior, contributing to the improvement of budget allocation strategies.
arXiv:2501.05280v2 Announce Type: replace Abstract: Involving people in energy systems planning can increase the legitimacy and socio-political feasibility of energy transitions. Participatory research in energy modelling offers the opportunity to engage with stakeholders in a comprehensive way, but is limited by how results can be generated and presented without imposing assumptions and discrete scenarios on the participants. To this end, we present a methodology and a framework, based on near-optimal modelling results, that can incorporate stakeholders in a holistic and engaging way. We confront stakeholders with a continuum of modelling-based energy system designs via an interactive interface allowing them to choose essentially any combination of components that meet the system requirements. Together with information on the implications of different technologies, it is possible to assess how participants prioritise different aspects in energy systems planning while also facilitating learning in an engaging and stimulating way. We showcase the methodology for the remote Arctic settlement of Longyearbyen and illustrate how participants deviate consistently from the cost optimum. At the same time, they manage to balance different priorities such as emissions, costs, and system vulnerability leading to a better understanding of the complexity and intertwined nature of decisions.
arXiv:2503.04020v1 Announce Type: cross Abstract: In traditional models of behavioral or opinion dynamics on social networks, researchers suppose that all interactions occur between pairs of individuals. However, in reality, social interactions also occur in groups of three or more individuals. A common way to incorporate such polyadic interactions is to study dynamical processes on hypergraphs. In a hypergraph, interactions can occur between any number of the individuals in a network. The Watts threshold model (WTM) is a well-known model of a simplistic social spreading process. Very recently, Chen et al. extended the WTM from dyadic networks (i.e., graphs) to polyadic networks (i.e., hypergraphs). In the present paper, we extend their discrete-time model to continuous time using approximate master equations (AMEs). By using AMEs, we are able to model the system with very high accuracy. We then reduce the high-dimensional AME system to a system of three coupled differential equations without any detectable loss of accuracy. This much lower-dimensional system is more computationally efficient to solve numerically and is also easier to interpret. We linearize the reduced AME system and calculate a cascade condition, which allows us to determine when a large spreading event occurs. We then apply our model to a social contact network of a French primary school and to a hypergraph of computer-science coauthorships. We find that the AME system is accurate in modelling the polyadic WTM on these empirical networks; however, we expect that future work that incorporates structural correlations between nearby nodes and groups into the model for the dynamics will lead to more accurate theory for real-world networks.
arXiv:2501.11219v1 Announce Type: cross Abstract: Mixed extension has played an important role in game theory, especially in the proof of the existence of Nash equilibria in strategic form games. Mixed extension can be regarded as continuous relaxation of a strategic form game. Recently, in repeated games, a class of behavior strategies, called zero-determinant strategies, was introduced. Zero-determinant strategies unilaterally enforce linear relations between payoffs, and are used to control payoffs of players. There are many attempts to extend zero-determinant strategies so as to apply them to broader situations. Here, we extend zero-determinant strategies to repeated games where action sets of players in stage game are continuously relaxed. We see that continuous relaxation broadens the range of possible zero-determinant strategies, compared to the original repeated games. Furthermore, we introduce a special type of zero-determinant strategies, called one-point zero-determinant strategies, which repeat only one continuously-relaxed action in all rounds. By investigating several examples, we show that some property of mixed-strategy Nash equilibria can be reinterpreted as a payoff-control property of one-point zero-determinant strategies.
arXiv:2204.08444v2 Announce Type: replace Abstract: Random network models, constrained to reproduce specific statistical features, are often used to represent and analyze network data and their mathematical descriptions. Chief among them, the configuration model constrains random networks by their degree distribution and is foundational to many areas of network science. However, configuration models and their variants are often selected based on intuition or mathematical and computational simplicity rather than on statistical evidence. To evaluate the quality of a network representation, we need to consider both the amount of information required to specify a random network model and the probability of recovering the original data when using the model as a generative process. To this end, we calculate the approximate size of network ensembles generated by the popular configuration model and its generalizations, including versions accounting for degree correlations and centrality layers. We then apply the minimum description length principle as a model selection criterion over the resulting nested family of configuration models. Using a dataset of over 100 networks from various domains, we find that the classic Configuration Model is generally preferred on networks with an average degree above ten, while a Layered Configuration Model constrained by a centrality metric offers the most compact representation of the majority of sparse networks.
arXiv:2501.11024v1 Announce Type: new Abstract: Networks significantly influence social, economic, and organizational outcomes, with centrality measures serving as crucial tools to capture the importance of individual nodes. This paper introduces Laplacian Eigenvector Centrality (LEC), a novel framework for network analysis based on spectral graph theory and the eigendecomposition of the Laplacian matrix. A distinctive feature of LEC is its adjustable parameter, the LEC order, which enables researchers to control and assess the scope of centrality measurement using the Laplacian spectrum. Using random graph models, LEC demonstrates robustness and scalability across diverse network structures. We connect LEC to equilibrium responses to external shocks in an economic model, showing how LEC quantifies agents' roles in attenuating shocks and facilitating coordinated responses through quadratic optimization. Finally, we apply LEC to the study of microfinance diffusion, illustrating how it complements classical centrality measures, such as eigenvector and Katz-Bonacich centralities, by capturing distinctive aspects of node positions within the network.
arXiv:2501.12198v1 Announce Type: new Abstract: This paper focuses on the opinion dynamics under the influence of manipulative agents. This type of agents is characterized by the fact that their opinions follow a trajectory that does not respond to the dynamics of the model, although it does influence the rest of the normal agents. Simulation has been implemented to study how one manipulative group modifies the natural dynamics of some opinion models of bounded confidence. It is studied what strategies based on the number of manipulative agents and their common opinion trajectory can be carried out by a manipulative group to influence normal agents and attract them to their opinions. In certain weighted models, some effects are observed in which normal agents move in the opposite direction to the manipulator group. Moreover, the conditions which ensure the influence of a manipulative group on a group of normal agents over time are also established for the Hegselmann-Krause model.
arXiv:2411.19094v2 Announce Type: replace-cross Abstract: Beautimeter is a new tool powered by generative pre-trained transformer (GPT) technology, designed to evaluate architectural and urban beauty. Rooted in Christopher Alexander's theory of centers, this work builds on the idea that all environments possess, to varying degrees, an innate sense of life. Alexander identified 15 fundamental properties, such as levels of scale and thick boundaries, that characterize living structure, which Beautimeter uses as a basis for its analysis. By integrating GPT's advanced natural language processing capabilities, Beautimeter assesses the extent to which a structure embodies these 15 properties, enabling a nuanced evaluation of architectural and urban aesthetics. Using ChatGPT, the tool helps users generate insights into the perceived beauty and coherence of spaces. We conducted a series of case studies, evaluating images of architectural and urban environments, as well as carpets, paintings, and other artifacts. The results demonstrate Beautimeter's effectiveness in analyzing aesthetic qualities across diverse contexts. Our findings suggest that by leveraging GPT technology, Beautimeter offers architects, urban planners, and designers a powerful tool to create spaces that resonate deeply with people. This paper also explores the implications of such technology for architecture and urban design, highlighting its potential to enhance both the design process and the assessment of built environments. Keywords: Living structure, structural beauty, Christopher Alexander, AI in Design, human centered design
arXiv:2501.10990v1 Announce Type: new Abstract: Citations in the scientific literature system do not simply reflect relationships between knowledge but are influenced by non-objective and societal factors. Citation bias, irresponsible citation, and citation manipulation are widespread and have become a serious and growing problem. However, it has been difficult to assess the consequences of mixing societal factors into the literature system because there was no observable literature system unmixed with societal factors for comparison. In this paper, we construct a mathematical theorem network, representing a logic-based and objective knowledge system, to address this problem. By comparing the mathematical theorem network and the scientific citation networks, we find that these two types of networks are significantly different in their structure and function. In particular, the reward function in citation networks is impaired: The scientific citation network fails to provide more recognition for more disruptive results, while the mathematical theorem network can achieve. We develop a network generation model that can create two types of links$\unicode{x2014}$logical and societal$\unicode{x2014}$to account for these differences. The model parameter $q$, which we call the human influence factor, can control the number of societal links and thus regulate the degree of mixing of societal factors in the networks. Under this design, the model successfully reproduces the differences among real networks. These results suggest that the presence of societal factors undermines the function of the scientific reward system. To improve the status quo, we advocate for reforming the reference list format in papers, urging journals to require authors to separately disclose logical references and social references.
arXiv:2501.07327v1 Announce Type: new Abstract: The advantages of temporal networks in capturing complex dynamics, such as diffusion and contagion, has led to breakthroughs in real world systems across numerous fields. In the case of human behavior, face-to-face interaction networks enable us to understand the dynamics of how communities emerge and evolve in time through the interactions, which is crucial in fields like epidemics, sociological studies and urban science. However, state-of-the-art datasets suffer from a number of drawbacks, such as short time-span for data collection and a small number of participants. Moreover, concerns arise for the participants' privacy and the data collection costs. Over the past years, many successful algorithms for static networks generation have been proposed, but they often do not tackle the social structure of interactions or their temporal aspect. In this work, we extend a recent network generation approach to capture the evolution of interactions between different communities. Our method labels nodes based on their community affiliation and constructs surrogate networks that reflect the interactions of the original temporal networks between nodes with different labels. This enables the generation of synthetic networks that replicate realistic behaviors. We validate our approach by comparing structural measures between the original and generated networks across multiple face-to-face interaction datasets.
arXiv:2311.11200v2 Announce Type: replace-cross Abstract: Scale-free networks play a fundamental role in the study of complex networks and various applied fields due to their ability to model a wide range of real-world systems. A key characteristic of these networks is their degree distribution, which often follows a power-law distribution, where the probability mass function is proportional to $x^{-\alpha}$, with $\alpha$ typically ranging between $2 < \alpha < 3$. In this paper, we introduce Bayesian inference methods to obtain more accurate estimates than those obtained using traditional methods, which often yield biased estimates, and precise credible intervals. Through a simulation study, we demonstrate that our approach provides nearly unbiased estimates for the scaling parameter, enhancing the reliability of inferences. We also evaluate new goodness-of-fit tests to improve the effectiveness of the Kolmogorov-Smirnov test, commonly used for this purpose. Our findings show that the Watson test offers superior power while maintaining a controlled type I error rate, enabling us to better determine whether data adheres to a power-law distribution. Finally, we propose a piecewise extension of this model to provide greater flexibility, evaluating the estimation and its goodness-of-fit features as well. In the complex networks field, this extension allows us to model the full degree distribution, instead of just focusing on the tail, as is commonly done. We demonstrate the utility of these novel methods through applications to two real-world datasets, showcasing their practical relevance and potential to advance the analysis of power-law behavior.
arXiv:2501.07193v1 Announce Type: cross Abstract: The lack of cooperation can easily result in inequality among members of a society, which provides an increasing gap between individual incomes. To tackle this issue, we introduce an incentive mechanism based on individual strategies and incomes, wherein a portion of the income from defectors is allocated to reward low-income cooperators, aiming to enhance cooperation by improving the equitable distribution of wealth across the entire population. Moreover, previous research has typically employed network structures or game mechanisms characterized by homogeneity. In this study, we present a network framework that more accurately reflects real-world conditions, where agents are engaged in multiple games, including prisoner's dilemma games in the top-layer and public good games in the down-layer networks. Within this framework, we introduce the concept of ``external coupling'' which connects agents across different networks as acquaintances, thereby facilitating access to shared datasets. Our results indicate that the combined positive effects of external coupling and incentive mechanism lead to optimal cooperation rates and lower Gini coefficients, demonstrating a negative correlation between cooperation and inequality. From a micro-level perspective, this phenomenon primarily arises from the regular network, whereas suboptimal outcomes are observed within the scale-free network. These observations help to give a deeper insight into the interplay between cooperation and wealth disparity in evolutionary games in large populations.
arXiv:2410.02987v2 Announce Type: cross Abstract: In this work, we study the effectiveness of employing archetypal aperiodic sequencing -- namely Fibonacci, Thue-Morse, and Rudin-Shapiro -- on the Parrondian effect. From a capital gain perspective, our results show that these series do yield a Parrondo's Paradox with the Thue-Morse based strategy outperforming not only the other two aperiodic strategies but benchmark Parrondian games with random and periodical ($AABBAABB\ldots$) switching as well. The least performing of the three aperiodic strategies is the Rudin-Shapiro. To elucidate the underlying causes of these results, we analyze the cross-correlation between the capital generated by the switching protocols and that of the isolated losing games. This analysis reveals that a strong anticorrelation with both isolated games is typically required to achieve a robust manifestation of Parrondo's effect. We also study the influence of the sequencing on the capital using the lacunarity and persistence measures. In general, we observe that the switching protocols tend to become less performing in terms of the capital as one increases the persistence and thus approaches the features of an isolated losing game. For the (log-)lacunarity, a property related to heterogeneity, we notice that for small persistence (less than 0.5) the performance increases with the lacunarity with a maximum around 0.4. In respect of this, our work shows that the optimization of a switching protocol is strongly dependent on a fine-tuning between persistence and heterogeneity.
arXiv:2501.07557v1 Announce Type: new Abstract: Music has always been central to human culture, reflecting and shaping traditions, emotions, and societal changes. Technological advancements have transformed how music is created and consumed, influencing tastes and the music itself. In this study, we use Network Science to analyze musical complexity. Drawing on $\approx20,000$ MIDI files across six macro-genres spanning nearly four centuries, we represent each composition as a weighted directed network to study its structural properties. Our results show that Classical and Jazz compositions have higher complexity and melodic diversity than recently developed genres. However, a temporal analysis reveals a trend toward simplification, with even Classical and Jazz nearing the complexity levels of modern genres. This study highlights how digital tools and streaming platforms shape musical evolution, fostering new genres while driving homogenization and simplicity.
arXiv:2408.00139v2 Announce Type: replace Abstract: The related concepts of partisan belief systems, issue alignment, and partisan sorting are central to our understanding of politics. These phenomena have been studied using measures of alignment between pairs of topics, or how much individuals' attitudes toward a topic reveal about their attitudes toward another topic. We introduce a higher-order measure that extends the assessment of alignment beyond pairs of topics by quantifying the amount of information individuals' opinions on one topic reveal about a set of topics simultaneously. Applying this approach to legislative voting behavior shows that parliamentary systems typically exhibit similar multiway alignment characteristics, but can change in response to shifting intergroup dynamics. In American National Election Studies surveys, our approach reveals a growing significance of party identification together with a consistent rise in multiway alignment over time.
arXiv:2501.07473v1 Announce Type: new Abstract: Political polarization, a key driver of social fragmentation, has drawn increasing attention for its role in shaping online and offline discourse. Despite significant efforts, accurately measuring polarization within ideological distributions remains a challenge. This study evaluates five widely used polarization measures, testing their strengths and weaknesses with synthetic datasets and a real-world case study on YouTube discussions during the 2020 U.S. Presidential Election. Building on these findings, we present a novel adaptation of Kleinberg's burst detection algorithm to improve mode detection in polarized distributions. By offering both a critical review and an innovative methodological tool, this work advances the analysis of ideological patterns in social media discourse.
arXiv:2501.07368v1 Announce Type: new Abstract: Social media play a key role in mobilizing collective action, holding the potential for studying the pathways that lead individuals to actively engage in addressing global challenges. However, quantitative research in this area has been limited by the absence of granular and large-scale ground truth about the level of participation in collective action among individual social media users. To address this limitation, we present a novel suite of text classifiers designed to identify expressions of participation in collective action from social media posts, in a topic-agnostic fashion. Grounded in the theoretical framework of social movement mobilization, our classification captures participation and categorizes it into four levels: recognizing collective issues, engaging in calls-to-action, expressing intention of action, and reporting active involvement. We constructed a labeled training dataset of Reddit comments through crowdsourcing, which we used to train BERT classifiers and fine-tune Llama3 models. Our findings show that smaller language models can reliably detect expressions of participation (weighted F1=0.71), and rival larger models in capturing nuanced levels of participation. By applying our methodology to Reddit, we illustrate its effectiveness as a robust tool for characterizing online communities in innovative ways compared to topic modeling, stance detection, and keyword-based methods. Our framework contributes to Computational Social Science research by providing a new source of reliable annotations useful for investigating the social dynamics of collective action.