File size: 57,224 Bytes
a32a103 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 |
---
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:50000
- loss:CosineSimilarityLoss
base_model: sentence-transformers/all-MiniLM-L6-v2
widget:
- source_sentence: 'An article on behavioral reinforcement learning:
Title: Cell-ŧype-specific responses to associative learning in the primary motor
cortex.
Abstract: The primary motor cortex (M1) is known to be a critical site for movement
initiation and motor learning. Surprisingly, it has also been shown to possess
reward-related activity, presumably to facilitate reward-based learning of new
movements. However, whether reward-related signals are represented among different
cell types in M1, and whether their response properties change after cue-reward
conditioning remains unclear. Here, we performed longitudinal in vivo two-photon
Ca2+ imaging to monitor the activity of different neuronal cell types in M1 while
mice engaged in a classical conditioning task. Our results demonstrate that most
of the major neuronal cell types in M1 showed robust but differential responses
to both the conditioned cue stimulus (CS) and reward, and their response properties
undergo cell-ŧype-specific modifications after associative learning. PV-INs’ responses
became more reliable to the CS, while VIP-INs’ responses became more reliable
to reward. Pyramidal neurons only showed robust responses to novel reward, and
they habituated to it after associative learning. Lastly, SOM-INs’ responses emerged
and became more reliable to both the CS and reward after conditioning. These observations
suggest that cue- and reward-related signals are preferentially represented among
different neuronal cell types in M1, and the distinct modifications they undergo
during associative learning could be essential in triggering different aspects
of local circuit reorganization in M1 during reward-based motor skill learning.'
sentences:
- 'An article on behavioral reinforcement learning:
Title: Learning to construct sentences in Spanish: A replication of the Weird
Word Order technique.
Abstract: In the present study, children''s early ability to organise words into
sentences was investigated using the Weird Word Order procedure with Spanish-speaking
children. Spanish is a language that allows for more flexibility in the positions
of subjects and objects, with respect to verbs, than other previously studied
languages (English, French, and Japanese). As in prior studies (Abbot-Smith et
al., 2001; Chang et al., 2009; Franck et al., 2011; Matthews et al., 2005, 2007),
we manipulated the relative frequency of verbs in training sessions with two age
groups (three-A nd four-year-old children). Results supported earlier findings
with regard to frequency: Children produced atypical word orders significantly
more often with infrequent verbs than with frequent verbs. The findings from the
present study support probabilistic learning models which allow higher levels
of flexibility and, in turn, oppose hypotheses that defend early access to advanced
grammatical knowledge.'
- 'An article on behavioral reinforcement learning:
Title: What are the computations of the cerebellum, the basal ganglia and the
cerebral cortex?.
Abstract: The classical notion that the cerebellum and the basal ganglia are dedicated
to motor control is under dispute given increasing evidence of their involvement
in non-motor functions. Is it then impossible to characterize the functions of
the cerebellum, the basal ganglia and the cerebral cortex in a simplistic manner?
This paper presents a novel view that their computational roles can be characterized
not by asking what are the ''goals'' of their computation, such as motor or sensory,
but by asking what are the ''methods'' of their computation, specifically, their
learning algorithms. There is currently enough anatomical, physiological, and
theoretical evidence to support the hypotheses that the cerebellum is a specialized
organism for supervised learning, the basal ganglia are for reinforcement learning,
and the cerebral cortex is for unsupervised learning.This paper investigates how
the learning modules specialized for these three kinds of learning can be assembled
into goal-oriented behaving systems. In general, supervised learning modules in
the cerebellum can be utilized as ''internal models'' of the environment. Reinforcement
learning modules in the basal ganglia enable action selection by an ''evaluation''
of environmental states. Unsupervised learning modules in the cerebral cortex
can provide statistically efficient representation of the states of the environment
and the behaving system. Two basic action selection architectures are shown, namely,
reactive action selection and predictive action selection. They can be implemented
within the anatomical constraint of the network linking these structures. Furthermore,
the use of the cerebellar supervised learning modules for state estimation, behavioral
simulation, and encapsulation of learned skill is considered. Finally, the usefulness
of such theoretical frameworks in interpreting brain imaging data is demonstrated
in the paradigm of procedural learning.'
- 'An article on behavioral reinforcement learning:
Title: Repeated decisions and attitudes to risk.
Abstract: In contrast to the underpinnings of expected utility, the experimental
pilot study results reported here suggest that current decisions may be influenced
both by past decisions and by the possibility of making decisions in the future.'
- source_sentence: 'An article on behavioral reinforcement learning:
Title: Sensory Evidence Accumulation Using Optic Flow in a Naturalistic Navigation
Task.
Abstract: Sensory evidence accumulation is considered a hallmark of decision-making
in noisy environments. Integration of sensory inputs has been traditionally studied
using passive stimuli, segregating perception from action. Lessons learned from
this approach, however, may not generalize to ethological behaviors like navigation,
where there is an active interplay between perception and action. We designed
a sensory-based sequential decision task in virtual reality in which humans and
monkeys navigated to a memorized location by integrating optic flow generated
by their own joystick movements. A major challenge in such closed-loop tasks is
that subjects’ actions will determine future sensory input, causing ambiguity
about whether they rely on sensory input rather than expectations based solely
on a learned model of the dynamics. To test whether subjects integrated optic
flow over time, we used three independent experimental manipulations, unpredictable
optic flow perturbations, which pushed subjects off their trajectory; gain manipulation
of the joystick controller, which changed the consequences of actions; and manipulation
of the optic flow density, which changed the information borne by sensory evidence.
Our results suggest that both macaques (male) and humans (female/male) relied
heavily on optic flow, thereby demonstrating a critical role for sensory evidence
accumulation during naturalistic action-perception closed-loop tasks.'
sentences:
- 'An article on behavioral reinforcement learning:
Title: The importance of decision making in causal learning from interventions.
Abstract: Recent research has focused on how interventions benefit causal learning.
This research suggests that the main benefit of interventions is in the temporal
and conditional probability information that interventions provide a learner.
But when one generates interventions, one must also decide what interventions
to generate. In three experiments, we investigated the importance of these decision
demands to causal learning. Experiment 1 demonstrated that learners were better
at learning causal models when they observed intervention data that they had generated,
as opposed to observing data generated by another learner. Experiment 2 demonstrated
the same effect between self-generated interventions and interventions learners
were forced to make. Experiment 3 demonstrated that when learners observed a sequence
of interventions such that the decision-making process that generated those interventions
was more readily available, learning was less impaired. These data suggest that
decision making may be an important part of causal learning from interventions.'
- 'An article on behavioral reinforcement learning:
Title: Region-specific effects of acute haloperidol in the human midbrain, striatum
and cortex.
Abstract: D2 autoreceptors provide an important regulatory mechanism of dopaminergic
neurotransmission. However, D2 receptors are also expressed as heteroreceptors
at postsynaptic membranes. The expression and the functional characteristics of
both, D2 auto- and heteroreceptors, differ between brain regions. Therefore, one
would expect that also the net response to a D2 antagonist, i.e. whether and to
what degree overall neural activity increases or decreases, varies across brain
areas. In the current study we systematically tested this hypothesis by parametrically
increasing haloperidol levels (placebo, 2 and 3 mg) in healthy volunteers and
measuring brain activity in the three major dopaminergic pathways. In particular,
activity was assessed using fMRI while participants performed a working memory
and a reinforcement learning task. Consistent with the hypothesis, across brain
regions activity parametrically in- and decreased. Moreover, even within the same
area there were function-specific concurrent de- and increases of activity, likely
caused by input from upstream dopaminergic regions. In the ventral striatum, for
instance, activity during reinforcement learning decreased for outcome processing
while prediction error related activity increased. In conclusion, the current
study highlights the intricacy of D2 neurotransmission which makes it difficult
to predict the function-specific net response of a given area to pharmacological
manipulations.'
- 'An article on behavioral reinforcement learning:
Title: Modeling dopaminergic and other processes involved in learning from reward
prediction error: Contributions from an individual differences perspective.
Abstract: Phasic firing changes of midbrain dopamine neurons have been widely
characterized as reflecting a reward prediction error (RPE). Major personality
traits (e.g., extraversion) have been linked to inter-individual variations in
dopaminergic neurotransmission. Consistent with these two claims, recent research
(Smillie et al., 2011; Cooper et al., 2014) found that extraverts exhibited larger
RPEs than introverts, as reflected in feedback related negativity (FRN) effects
in EEG recordings. Using an established, biologically-localized RPE computational
model, we successfully simulated dopaminergic cell firing changes which are thought
to modulate the FRN. We introduced simulated individual differences into the model:
parameters were systematically varied, with stable values for each simulated individual.
We explored whether a model parameter might be responsible for the observed covariance
between extraversion and the FRN changes in real data, and argued that a parameter
is a plausible source of such covariance if parameter variance, across simulated
individuals, correlated almost perfectly with the size of the simulated dopaminergic
FRN modulation, and created as much variance as possible in this simulated output.
Several model parameters met these criteria, while others did not. In particular,
variations in the strength of connections carrying excitatory reward drive inputs
to midbrain dopaminergic cells were considered plausible candidates, along with
variations in a parameter which scales the effects of dopamine cell firing bursts
on synaptic modification in ventral striatum. We suggest possible neurotransmitter
mechanisms underpinning these model parameters. Finally, the limitations and possible
extensions of our general approach are discussed.'
- source_sentence: 'An article on behavioral reinforcement learning:
Title: Pigeons'' use of cues in a repeated five-trial-sequence, single-reversal
task.
Abstract: We studied behavioral flexibility, or the ability to modify one''s behavior
in accordance with the changing environment, in pigeons using a reversal-learning
paradigm. In two experiments, each session consisted of a series of five-trial
sequences involving a simple simultaneous color discrimination in which a reversal
could occur during each sequence. The ideal strategy would be to start each sequence
with a choice of S1 (the first correct stimulus) until it was no longer correct,
and then to switch to S2 (the second correct stimulus), thus utilizing cues provided
by local reinforcement (feedback from the preceding trial). In both experiments,
subjects showed little evidence of using local reinforcement cues, but instead
used the mean probabilities of reinforcement for S1 and S2 on each trial within
each sequence. That is, subjects showed remarkably similar behavior, regardless
of where (or, in Exp. 2, whether) a reversal occurred during a given sequence.
Therefore, subjects appeared to be relatively insensitive to the consequences
of responses (local feedback) and were not able to maximize reinforcement. The
fact that pigeons did not use the more optimal feedback afforded by recent reinforcement
contingencies to maximize their reinforcement has implications for their use of
flexible response strategies under reversal-learning conditions.'
sentences:
- 'An article on behavioral reinforcement learning:
Title: Behavioral and circuit basis of sucrose rejection by drosophila females
in a simple decision-making task.
Abstract: Drosophila melanogaster egg-laying site selection offers a genetic model
to study a simple form of value-based decision. We have previously shown that
Drosophila females consistently reject a sucrose-containing substrate and choose
a plain (sucrose-free) substrate for egg laying in our sucrose versus plain decision
assay. However, either substrate is accepted when it is the sole option. Here
we describe the neural mechanism that underlies females’ sucrose rejection in
our sucrose versus plain assay. First, we demonstrate that females explored the
sucrose substrate frequently before most egg-laying events, suggesting that they
actively suppress laying eggs on the sucrose substrate as opposed to avoiding
visits to it. Second, we show that activating a specific subset of DA neurons
triggered a preference for laying eggs on the sucrose substrate over the plain
one, suggesting that activating these DA neurons can increase the value of the
sucrose substrate for egg laying. Third, we demonstrate that neither ablating
nor inhibiting the mushroom body (MB), a known Drosophila learning and decision
center, affected females’ egg-laying preferences in our sucrose versus plain assay,
suggesting that MB does not mediate this specific decision-making task.Wepropose
that the value of a sucrose substrate— as an egg-laying option—can be adjusted
by the activities of a specific DA circuit. Once the sucrose substrate is determined
to be the lesser valued option, females execute their decision to reject this
inferior substrate not by stopping their visits to it, but by actively suppressing
their egg-laying motor program during their visits.'
- 'An article on behavioral reinforcement learning:
Title: Choice in experiential learning: True preferences or experimental artifacts?.
Abstract: The rate of selecting different options in the decisions-from-feedback
paradigm is commonly used to measure preferences resulting from experiential learning.
While convergence to a single option increases with experience, some variance
in choice remains even when options are static and offer fixed rewards. Employing
a decisions-from-feedback paradigm followed by a policy-setting task, we examined
whether the observed variance in choice is driven by factors related to the paradigm
itself: Continued exploration (e.g., believing options are non-stationary) or
exploitation of perceived outcome patterns (i.e., a belief that sequential choices
are not independent). Across two studies, participants showed variance in their
choices, which was related (i.e., proportional) to the policies they set. In addition,
in Study 2, participants'' reported under-confidence was associated with the amount
of choice variance in later choices and policies. These results suggest that variance
in choice is better explained by participants lacking confidence in knowing which
option is better, rather than methodological artifacts (i.e., exploration or failures
to recognize outcome independence). As such, the current studies provide evidence
for the decisions-from-feedback paradigm''s validity as a behavioral research
method for assessing learned preferences.'
- 'An article on behavioral reinforcement learning:
Title: Impaired savings despite intact initial learning of motor adaptation in
Parkinson''s disease.
Abstract: In motor adaptation, the occurrence of savings (faster relearning of
a previously learned motor adaptation task) has been explained in terms of operant
reinforcement learning (Huang et al. in Neuron 70(4):787-801, 2011), which is
thought to associate an adapted motor command with outcome success during repeated
execution of the adapted movement. There is some evidence for deficient savings
in Parkinson''s Disease (PD), which might result from deficient operant reinforcement
processes. However, this evidence is compromised by limited adaptation training
during initial learning and by multi-target adaptation, which reduces the number
of reinforced movement repetitions for each target. Here, we examined savings
in PD patients and controls following overlearning with a single target. PD patients
showed less savings than controls after successive adaptation and deadaptation
blocks within the same test session, as well as less savings across test sessions
separated by a 24-h delay. It is argued that impaired blunted dopaminergic signals
in PD impairs the modulation of dopaminergic signals to the motor cortex in response
to rewarding motor outcomes, thus impairing the association of the adapted motor
command with rewarding motor outcomes. Consequently, the previously adapted motor
command is not preferentially selected during relearning, and savings is impaired.'
- source_sentence: 'An article on behavioral reinforcement learning:
Title: Altered cingulate sub-region activation accounts for task-related dissociation
in ERN amplitude as a function of obsessive-compulsive symptoms.
Abstract: Larger error-related negativities (ERNs) have been consistently found
in obsessive-compulsive disorder (OCD) patients, and are thought to reflect the
activities of a hyperactive cortico-striatal circuit during action monitoring.
We previously observed that obsessive-compulsive (OC) symptomatic students (non-patients)
have larger ERNs during errors in a response competition task, yet smaller ERNs
in a reinforcement learning task. The finding of a task-specific dissociation
suggests that distinct yet partially overlapping medio-frontal systems underlie
the ERN in different tasks, and that OC symptoms are associated with functional
differences in these systems. Here, we used EEG source localization to identify
why OC symptoms are associated with hyperactive ERNs to errors yet hypoactive
ERNs when selecting maladaptive actions. At rest, OC symptomatology predicted
greater activity in rostral anterior cingulate cortex (rACC) and lower activity
in dorsal anterior cingulate cortex (dACC). When compared to a group with low
OC symptom scores, the high OC group had greater rACC reactivity during errors
in the response competition task and less deactivation of dACC activity during
errors in the reinforcement learning task. The degree of activation in these areas
correlated with ERN amplitudes during both tasks in the high OC group, but not
in the low group. Interactive anterior cingulate cortex (ACC) systems associated
avoidance of maladaptive actions were intact in the high OC group, but were related
to poorer performance on a third task: probabilistic reversal learning. These
novel findings link both tonic and phasic activities in the ACC to action monitoring
alterations, including dissociation in performance deficits, in OC symptomatic
participants.'
sentences:
- 'An article on behavioral reinforcement learning:
Title: The Stroop Effect: Why Proportion Congruent Has Nothing to Do With Congruency
and Everything to Do With Contingency.
Abstract: The item-specific proportion congruent (ISPC) effect refers to the observation
that the Stroop effect is larger for words that are presented mostly in congruent
colors (e.g., BLUE presented 75% of the time in blue) and smaller for words that
are presented mostly in a given incongruent color (e.g., YELLOW presented 75%
of the time in orange). One account of the ISPC effect, the modulation hypothesis,
is that participants modulate attention based on the identity of the word (i.e.,
participants allow the word to influence responding when it is presented mostly
in its congruent color). Another account, the contingency hypothesis, is that
participants use the word to predict the response that they will need to make
(e.g., if the word is YELLOW, then the response is probably "orange"). Reanalyses
of data from L. L. Jacoby, D. S. Lindsay, and S. Hessels (2003), along with results
from new experiments, are inconsistent with the modulation hypothesis but entirely
consistent with the contingency hypothesis. A response threshold mechanism that
uses contingency information provides a sufficient account of the data.'
- 'An article on behavioral reinforcement learning:
Title: D-cycloserine facilitates socially reinforced learning in an animal model
relevant to autism spectrum disorders.
Abstract: There are no drugs that specifically target the social deficits of autism
spectrum disorders (ASD). This may be due to a lack of behavioral paradigms in
animal models relevant to ASD. Partner preference formation in the prairie vole
represents a social cognitive process involving socially reinforced learning.
D-cycloserine (DCS) is a cognitive enhancer that acts at the N-methyl-D-aspartate
receptor to promote learning. If DCS enhances socially reinforced learning in
the partner preference paradigm, it may be useful in combination with behavioral
therapies for enhancing social functioning in ASD. Female prairie and meadow voles
were given DCS either peripherally or directly into one of three brain regions:
nucleus accumbens, amygdala, or caudate putamen. Subjects were then cohabited
with a male vole under conditions that do not typically yield a partner preference.
The development of a preference for that stimulus male vole over a novel male
vole was assessed using a partner preference test. A low dose of DCS administered
peripherally enhanced preference formation in prairie voles but not meadow voles
under conditions in which it would not otherwise occur. These effects were replicated
in prairie voles by microinfusions of DCS into the nucleus accumbens, which is
involved in reinforcement learning, and the amygdala, which is involved in social
information processing. Partner preference in the prairie vole may provide a behavioral
paradigm with face, construct, and predictive validity for identifying prosocial
pharmacotherapeutics. D-cycloserine may be a viable treatment strategy for social
deficits of ASD when paired with social behavioral therapy.'
- 'An article on behavioral reinforcement learning:
Title: Pseudodiagnosticity Revisited.
Abstract: In the psychology of reasoning and judgment, the pseudodiagnosticity
task has been a major tool for the empirical investigation of people''s ability
to search for diagnostic information. A novel normative analysis of this experimental
paradigm is presented, by which the participants'' prevailing responses turn out
not to support the generally accepted existence of a reasoning bias. The conclusions
drawn do not rest on pragmatic concerns suggesting alleged divergences between
the experimenter''s and participants'' reading of the task. They only rely, instead,
on the demonstration that observed behavior largely conforms to optimal utility
maximizing information search strategies for standard variants of the pseudodiagnosticity
paradigm that have been investigated so far. It is argued that the experimental
results obtained, contrary to what has recurrently been claimed, have failed to
discriminate between normative and nonnormative accounts of behavior. More general
implications of the analysis presented for past and future research on human information
search behavior and diagnostic reasoning are discussed.'
- source_sentence: 'An article on behavioral reinforcement learning:
Title: Confidence and the description–experience distinction.
Abstract: In this paper, we extend the literature on the description–experience
gap in risky choices by focusing on how the mode of learning—through description
or experience—affects confidence. Specifically, we explore how learning through
description or experience affects confidence in (1) the information gathered to
make a decision and (2) the resulting choice. In two preregistered experiments
we tested whether there was a description–experience gap in both dimensions of
confidence. Learning from description was associated with higher confidence—both
in the information gathered and in the choice made—than was learning from experience.
In a third preregistered experiment, we examined the effect of sample size on
confidence in decisions from experience. Contrary to the normative view that larger
samples foster confidence in statistical inference, we observed that more experience
led to less confidence. This observation is reminiscent of recent theories of
deliberate ignorance, which highlight the adaptive benefits of deliberately limiting
information search.'
sentences:
- 'An article on behavioral reinforcement learning:
Title: Episodic memories predict adaptive Value-Based Decision-Making.
Abstract: Prior research illustrates that memory can guide Value-Based Decision-Making.
For example, previous work has implicated both working memory and procedural memory
(i.e., reinforcement learning) in guiding choice. However, other types of memories,
such as episodic memory, may also influence Decision-Making. Here we test the
role for episodic Memory-Specifically item versus associative Memory-In supporting
Value-Based choice. Participants completed a task where they first learned the
value associated with trial unique lotteries. After a short delay, they completed
a Decision-Making task where they could choose to reengage with previously encountered
lotteries, or new never before seen lotteries. Finally, participants completed
a surprise memory test for the lotteries and their associated values. Results
indicate that participants chose to reengage more often with lotteries that resulted
in high versus low rewards. Critically, participants not only formed detailed,
associative memories for the reward values coupled with individual lotteries,
but also exhibited adaptive Decision-Making only when they had intact associative
memory. We further found that the relationship between adaptive choice and associative
memory generalized to more complex, ecologically valid choice behavior, such as
social decisionmaking. However, individuals more strongly encode experiences of
social Violations-Such as being treated unfairly, suggesting a bias for how individuals
form associative memories within social contexts. Together, these findings provide
an important integration of episodic memory and Decision-Making literatures to
better understand key mechanisms supporting adaptive behavior.'
- 'An article on behavioral reinforcement learning:
Title: How (in)variant are subjective representations of described and experienced
risk and rewards?.
Abstract: Decisions under risk have been shown to differ depending on whether
information on outcomes and probabilities is gleaned from symbolic descriptions
or gathered through experience. To some extent, this description–experience gap
is due to sampling error in experience-based choice. Analyses with cumulative
prospect theory (CPT), investigating to what extent the gap is also driven by
differences in people''s subjective representations of outcome and probability
information (taking into account sampling error), have produced mixed results.
We improve on previous analyses of description-based and experience-based choices
by taking advantage of both a within-subjects design and a hierarchical Bayesian
implementation of CPT. This approach allows us to capture both the differences
and the within-person stability of individuals’ subjective representations across
the two modes of learning about choice options. Relative to decisions from description,
decisions from experience showed reduced sensitivity to probabilities and increased
sensitivity to outcomes. For some CPT parameters, individual differences were
relatively stable across modes of learning. Our results suggest that outcome and
probability information translate into systematically different subjective representations
in description- versus experience-based choice. At the same time, both types of
decisions seem to tap into the same individual-level regularities.'
- 'An article on behavioral reinforcement learning:
Title: Do narcissists make better decisions? An investigation of narcissism and
dynamic decision-making performance.
Abstract: We investigated whether narcissism affected dynamic decision-making
performance in the presence and absence of misleading information. Performance
was examined in a two-choice dynamic decision-making task where the optimal strategy
was to forego an option providing larger immediate rewards in favor of an option
that led to larger delayed rewards. Information regarding foregone rewards from
the alternate option was presented or withheld to bias participants toward the
sub-optimal choice. The results demonstrated that individuals high in narcissistic
traits performed comparably to low narcissism individuals when foregone reward
information was absent, but high narcissism individuals outperformed individuals
low in narcissistic traits when misleading information was presented. The advantage
for participants high in narcissistic traits was strongest within males, and,
overall, males outperformed females when foregone rewards were present. While
prior research emphasizes narcissists'' decision-making deficits, our findings
provide evidence that individuals high in narcissistic traits excel at decision-making
tasks that involve disregarding ambiguous information and focusing on the long-term
utility of each option. Their superior ability at filtering out misleading information
may reflect an effort to maintain their self-view or avoid ego threat.'
pipeline_tag: sentence-similarity
library_name: sentence-transformers
---
# SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
## Model Details
### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
- **Maximum Sequence Length:** 256 tokens
- **Output Dimensionality:** 384 dimensions
- **Similarity Function:** Cosine Similarity
<!-- - **Training Dataset:** Unknown -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
### Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Normalize()
)
```
## Usage
### Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
```bash
pip install -U sentence-transformers
```
Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("dwulff/minilm-brl")
# Run inference
sentences = [
'An article on behavioral reinforcement learning:\n\nTitle: Confidence and the description–experience distinction.\nAbstract: In this paper, we extend the literature on the description–experience gap in risky choices by focusing on how the mode of learning—through description or experience—affects confidence. Specifically, we explore how learning through description or experience affects confidence in (1) the information gathered to make a decision and (2) the resulting choice. In two preregistered experiments we tested whether there was a description–experience gap in both dimensions of confidence. Learning from description was associated with higher confidence—both in the information gathered and in the choice made—than was learning from experience. In a third preregistered experiment, we examined the effect of sample size on confidence in decisions from experience. Contrary to the normative view that larger samples foster confidence in statistical inference, we observed that more experience led to less confidence. This observation is reminiscent of recent theories of deliberate ignorance, which highlight the adaptive benefits of deliberately limiting information search.',
"An article on behavioral reinforcement learning:\n\nTitle: How (in)variant are subjective representations of described and experienced risk and rewards?.\nAbstract: Decisions under risk have been shown to differ depending on whether information on outcomes and probabilities is gleaned from symbolic descriptions or gathered through experience. To some extent, this description–experience gap is due to sampling error in experience-based choice. Analyses with cumulative prospect theory (CPT), investigating to what extent the gap is also driven by differences in people's subjective representations of outcome and probability information (taking into account sampling error), have produced mixed results. We improve on previous analyses of description-based and experience-based choices by taking advantage of both a within-subjects design and a hierarchical Bayesian implementation of CPT. This approach allows us to capture both the differences and the within-person stability of individuals’ subjective representations across the two modes of learning about choice options. Relative to decisions from description, decisions from experience showed reduced sensitivity to probabilities and increased sensitivity to outcomes. For some CPT parameters, individual differences were relatively stable across modes of learning. Our results suggest that outcome and probability information translate into systematically different subjective representations in description- versus experience-based choice. At the same time, both types of decisions seem to tap into the same individual-level regularities.",
"An article on behavioral reinforcement learning:\n\nTitle: Do narcissists make better decisions? An investigation of narcissism and dynamic decision-making performance.\nAbstract: We investigated whether narcissism affected dynamic decision-making performance in the presence and absence of misleading information. Performance was examined in a two-choice dynamic decision-making task where the optimal strategy was to forego an option providing larger immediate rewards in favor of an option that led to larger delayed rewards. Information regarding foregone rewards from the alternate option was presented or withheld to bias participants toward the sub-optimal choice. The results demonstrated that individuals high in narcissistic traits performed comparably to low narcissism individuals when foregone reward information was absent, but high narcissism individuals outperformed individuals low in narcissistic traits when misleading information was presented. The advantage for participants high in narcissistic traits was strongest within males, and, overall, males outperformed females when foregone rewards were present. While prior research emphasizes narcissists' decision-making deficits, our findings provide evidence that individuals high in narcissistic traits excel at decision-making tasks that involve disregarding ambiguous information and focusing on the long-term utility of each option. Their superior ability at filtering out misleading information may reflect an effort to maintain their self-view or avoid ego threat.",
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 384]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
```
<!--
### Direct Usage (Transformers)
<details><summary>Click to see the direct usage in Transformers</summary>
</details>
-->
<!--
### Downstream Usage (Sentence Transformers)
You can finetune this model on your own dataset.
<details><summary>Click to expand</summary>
</details>
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Dataset
#### Unnamed Dataset
* Size: 50,000 training samples
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
* Approximate statistics based on the first 1000 samples:
| | sentence_0 | sentence_1 | label |
|:--------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:---------------------------------------------------------------|
| type | string | string | float |
| details | <ul><li>min: 102 tokens</li><li>mean: 237.66 tokens</li><li>max: 256 tokens</li></ul> | <ul><li>min: 61 tokens</li><li>mean: 227.84 tokens</li><li>max: 256 tokens</li></ul> | <ul><li>min: 0.0</li><li>mean: 0.17</li><li>max: 0.9</li></ul> |
* Samples:
| sentence_0 | sentence_1 | label |
|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------|
| <code>An article on behavioral reinforcement learning:<br><br>Title: Working memory and response selection: A computational account of interactions among cortico-basalganglio-thalamic loops.<br>Abstract: Cortico-basalganglio-thalamic loops are involved in both cognitive processes and motor control. We present a biologically meaningful computational model of how these loops contribute to the organization of working memory and the development of response behavior. Via reinforcement learning in basal ganglia, the model develops flexible control of working memory within prefrontal loops and achieves selection of appropriate responses based on working memory content and visual stimulation within a motor loop. We show that both working memory control and response selection can evolve within parallel and interacting cortico-basalganglio-thalamic loops by Hebbian and three-factor learning rules. Furthermore, the model gives a coherent explanation for how complex strategies of working memory control and respo...</code> | <code>An article on behavioral reinforcement learning:<br><br>Title: The role of basal ganglia in reinforcement learning and imprinting in domestic chicks.<br>Abstract: Effects of bilateral kainate lesions of telencephalic basal ganglia (lobus parolfactorius, LPO) were examined in domestic chicks. In the imprinting paradigm, where chicks learned to selectively approach a moving object without any explicitly associated reward, both the pre- and post-training lesions were without effects. On the other hand, in the water-reinforced pecking task, pre-training lesions of LPO severely impaired immediate reinforcement as well as formation of the association memory. However, post-training LPO lesions did not cause amnesia, and chicks selectively pecked at the reinforced color. The LPO could thus be involved specifically in the evaluation of present rewards and the instantaneous reinforcement of pecking, but not in the execution of selective behavior based on a memorized color cue.</code> | <code>0.5</code> |
| <code>An article on behavioral reinforcement learning:<br><br>Title: Exploration Disrupts Choice-Predictive Signals and Alters Dynamics in Prefrontal Cortex.<br>Abstract: In uncertain environments, decision-makers must balance two goals: they must “exploit” rewarding options but also “explore” in order to discover rewarding alternatives. Exploring and exploiting necessarily change how the brain responds to identical stimuli, but little is known about how these states, and transitions between them, change how the brain transforms sensory information into action. To address this question, we recorded neural activity in a prefrontal sensorimotor area while monkeys naturally switched between exploring and exploiting rewarding options. We found that exploration profoundly reduced spatially selective, choice-predictive activity in single neurons and delayed choice-predictive population dynamics. At the same time, reward learning was increased in brain and behavior. These results indicate that exploration i...</code> | <code>An article on behavioral reinforcement learning:<br><br>Title: Counterfactual choice and learning in a Neural Network centered on human lateral frontopolar cortex.<br>Abstract: Decision making and learning in a real-world context require organisms to track not only the choices they make and the outcomes that follow but also other untaken, or counterfactual, choices and their outcomes. Although the neural system responsible for tracking the value of choices actually taken is increasingly well understood, whether a neural system tracks counterfactual information is currently unclear. Using a three-alternative decision-making task, a Bayesian reinforcement-learning algorithm, and fMRI, we investigated the coding of counterfactual choices and prediction errors in the human brain. Rather than representing evidence favoring multiple counterfactual choices, lateral frontal polar cortex (lFPC), dorsomedial frontal cortex (DMFC), and posteromedial cortex (PMC) encode the reward-based evidence favoring t...</code> | <code>0.5</code> |
| <code>An article on behavioral reinforcement learning:<br><br>Title: Electrophysiological signatures of visual statistical learning in 3-month-old infants at familial and low risk for autism spectrum disorder.<br>Abstract: Visual statistical learning (VSL) refers to the ability to extract associations and conditional probabilities within the visual environment. It may serve as a precursor to cognitive and social communication development. Quantifying VSL in infants at familial risk (FR) for Autism Spectrum Disorder (ASD) provides opportunities to understand how genetic predisposition can influence early learning processes which may, in turn, lay a foundation for cognitive and social communication delays. We examined electroencephalography (EEG) signatures of VSL in 3-month-old infants, examining whether EEG correlates of VSL differentiated FR from low-risk (LR) infants. In an exploratory analysis, we then examined whether EEG correlates of VSL at 3 months relate to cognitive function and ASD symptoms...</code> | <code>An article on behavioral reinforcement learning:<br><br>Title: Reduced nucleus accumbens reactivity and adolescent depression following early-life stress.<br>Abstract: Depression is a common outcome for those having experienced early-life stress (ELS). For those individuals, depression typically increases during adolescence and appears to endure into adulthood, suggesting alterations in the development of brain systems involved in depression. Developmentally, the nucleus accumbens (NAcc), a limbic structure associated with reward learning and motivation, typically undergoes dramatic functional change during adolescence; therefore, age-related changes in NAcc function may underlie increases in depression in adolescence following ELS. The current study examined the effects of ELS in 38 previously institutionalized children and adolescents in comparison to a group of 31 youths without a history of ELS. Consistent with previous research, the findings showed that depression was higher in adolescents...</code> | <code>0.0</code> |
* Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
```json
{
"loss_fct": "torch.nn.modules.loss.MSELoss"
}
```
### Training Hyperparameters
#### Non-Default Hyperparameters
- `per_device_train_batch_size`: 64
- `per_device_eval_batch_size`: 64
- `num_train_epochs`: 5
- `multi_dataset_batch_sampler`: round_robin
#### All Hyperparameters
<details><summary>Click to expand</summary>
- `overwrite_output_dir`: False
- `do_predict`: False
- `eval_strategy`: no
- `prediction_loss_only`: True
- `per_device_train_batch_size`: 64
- `per_device_eval_batch_size`: 64
- `per_gpu_train_batch_size`: None
- `per_gpu_eval_batch_size`: None
- `gradient_accumulation_steps`: 1
- `eval_accumulation_steps`: None
- `torch_empty_cache_steps`: None
- `learning_rate`: 5e-05
- `weight_decay`: 0.0
- `adam_beta1`: 0.9
- `adam_beta2`: 0.999
- `adam_epsilon`: 1e-08
- `max_grad_norm`: 1
- `num_train_epochs`: 5
- `max_steps`: -1
- `lr_scheduler_type`: linear
- `lr_scheduler_kwargs`: {}
- `warmup_ratio`: 0.0
- `warmup_steps`: 0
- `log_level`: passive
- `log_level_replica`: warning
- `log_on_each_node`: True
- `logging_nan_inf_filter`: True
- `save_safetensors`: True
- `save_on_each_node`: False
- `save_only_model`: False
- `restore_callback_states_from_checkpoint`: False
- `no_cuda`: False
- `use_cpu`: False
- `use_mps_device`: False
- `seed`: 42
- `data_seed`: None
- `jit_mode_eval`: False
- `use_ipex`: False
- `bf16`: False
- `fp16`: False
- `fp16_opt_level`: O1
- `half_precision_backend`: auto
- `bf16_full_eval`: False
- `fp16_full_eval`: False
- `tf32`: None
- `local_rank`: 0
- `ddp_backend`: None
- `tpu_num_cores`: None
- `tpu_metrics_debug`: False
- `debug`: []
- `dataloader_drop_last`: False
- `dataloader_num_workers`: 0
- `dataloader_prefetch_factor`: None
- `past_index`: -1
- `disable_tqdm`: False
- `remove_unused_columns`: True
- `label_names`: None
- `load_best_model_at_end`: False
- `ignore_data_skip`: False
- `fsdp`: []
- `fsdp_min_num_params`: 0
- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
- `tp_size`: 0
- `fsdp_transformer_layer_cls_to_wrap`: None
- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
- `deepspeed`: None
- `label_smoothing_factor`: 0.0
- `optim`: adamw_torch
- `optim_args`: None
- `adafactor`: False
- `group_by_length`: False
- `length_column_name`: length
- `ddp_find_unused_parameters`: None
- `ddp_bucket_cap_mb`: None
- `ddp_broadcast_buffers`: False
- `dataloader_pin_memory`: True
- `dataloader_persistent_workers`: False
- `skip_memory_metrics`: True
- `use_legacy_prediction_loop`: False
- `push_to_hub`: False
- `resume_from_checkpoint`: None
- `hub_model_id`: None
- `hub_strategy`: every_save
- `hub_private_repo`: None
- `hub_always_push`: False
- `gradient_checkpointing`: False
- `gradient_checkpointing_kwargs`: None
- `include_inputs_for_metrics`: False
- `include_for_metrics`: []
- `eval_do_concat_batches`: True
- `fp16_backend`: auto
- `push_to_hub_model_id`: None
- `push_to_hub_organization`: None
- `mp_parameters`:
- `auto_find_batch_size`: False
- `full_determinism`: False
- `torchdynamo`: None
- `ray_scope`: last
- `ddp_timeout`: 1800
- `torch_compile`: False
- `torch_compile_backend`: None
- `torch_compile_mode`: None
- `dispatch_batches`: None
- `split_batches`: None
- `include_tokens_per_second`: False
- `include_num_input_tokens_seen`: False
- `neftune_noise_alpha`: None
- `optim_target_modules`: None
- `batch_eval_metrics`: False
- `eval_on_start`: False
- `use_liger_kernel`: False
- `eval_use_gather_object`: False
- `average_tokens_across_devices`: False
- `prompts`: None
- `batch_sampler`: batch_sampler
- `multi_dataset_batch_sampler`: round_robin
</details>
### Training Logs
| Epoch | Step | Training Loss |
|:------:|:----:|:-------------:|
| 0.6394 | 500 | 0.0179 |
| 1.2788 | 1000 | 0.0124 |
| 1.9182 | 1500 | 0.0107 |
| 2.5575 | 2000 | 0.0092 |
| 3.1969 | 2500 | 0.0086 |
| 3.8363 | 3000 | 0.0078 |
| 4.4757 | 3500 | 0.0073 |
### Framework Versions
- Python: 3.13.2
- Sentence Transformers: 4.0.2
- Transformers: 4.50.0.dev0
- PyTorch: 2.6.0
- Accelerate: 1.5.2
- Datasets: 3.5.0
- Tokenizers: 0.21.1
## Citation
### BibTeX
#### Sentence Transformers
```bibtex
@inproceedings{reimers-2019-sentence-bert,
title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
author = "Reimers, Nils and Gurevych, Iryna",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
month = "11",
year = "2019",
publisher = "Association for Computational Linguistics",
url = "https://arxiv.org/abs/1908.10084",
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> |