16 research outputs found

    Comparison of Subjective Responses to Oral and Intravenous Alcohol Administration under Similar Systemic Exposures

    Get PDF
    Objective To test whether an individual's subjective responses to alcohol are similar when the breath alcohol concentration (BrAC) trajectory resulting from oral administration is matched by intravenous administration. Background Individuals perceive the effects of alcohol differently, and the variation is commonly used in research assessing the risk for developing an alcohol use disorder. Such research is supported by both oral and intravenous alcohol administration techniques, and any differences attributable to the route employed should be understood. Methods We conducted a 2‐session, within‐subject study in 44 young adult, healthy, non‐dependent drinkers (22 females and 22 males). In the first session, subjects ingested a dose of alcohol which was individually calculated, on the basis of total body water, to yield a peak BrAC near 80 mg/dl, and the resulting BrAC trajectory was recorded. A few days later, subjects received an intravenous alcohol infusion rate profile, pre‐computed to replicate each individual's oral alcohol BrAC trajectory. In both sessions, we assessed 4 subjective responses to alcohol: SEDATION, SIMULATION, INTOXICATION, and HIGH; at baseline and frequently for 4 hours. We compared the individuals’ baseline‐corrected responses at peak BrAC and at half‐peak BrAC on both the ascending and descending limbs. We also computed and compared Pearson‐product moment correlations of responses by route of administration, the Mellanby measure of acute adaptation to alcohol, and the area under the entire response curve for each subjective response. Results No significant differences in any measure could be attributed to the route of alcohol administration. Eleven of 12 response comparisons were significantly correlated across the routes of alcohol administration, with 9 surviving correction for multiple measures, as did the Mellanby effect and area under the response curve correlations. Conclusion The route of alcohol administration has a minimal effect on subjective responses to alcohol when an individual's BrAC exposure profiles are similar

    Intravenous Alcohol Self-Administration in the P Rat

    Get PDF
    Alcohol consumption produces a complex array of effects that can be divided into two types: the explicit pharmacological effects of ethanol (which can be temporally separate from time of intake) and the more temporally “relevant” effects (primarily olfactory and taste) that bridge the time from intake to onset of the pharmacological effects. Intravenous (IV) self-administration of ethanol limits the confounding “non-pharmacological” effects associated with oral consumption, allows for controlled and precise dosing, and bypasses first order absorption kinetics, allowing for more direct and better-controlled assessment of alcohol’s effect on the brain. IV ethanol self-administration has been reliably demonstrated in mouse and human experimental models; however, models of IV self-administration have been historically problematic in the rat. An operant multiple-schedule study design was used to elucidate the role of each component of a compound IV-ethanol plus oral-sucrose reinforcer. Male alcohol-preferring P rats had free access to both food and water during all IV self-administration sessions. Animals were trained to press a lever for orally delivered 1% sucrose (1S) on a fixed ratio 4 schedule, and then surgically implanted with an indwelling jugular catheter. Animals were then trained to respond on a multiple FR4-FR4 schedule composed of alternating 2.5-min components across 30-min sessions. For the multiple schedule, two components were used: an oral 1S only and an oral 1S plus IV 20% ethanol (25 mg/kg/injection). Average total ethanol intake was 0.47 ± 0.04 g/kg. We found significantly higher earning of sucrose-only reinforcers and greater sucrose-lever error responding relative to the compound oral-sucrose plus IV-ethanol reinforcer. These response patterns suggest that sucrose, not ethanol, was responsible for driving overall responding. The work with a compound IV ethanol-oral sucrose reinforcer presented here suggests that the existing intravenous ethanol self-administration methodology cannot overcome the aversive properties of ethanol via this route in the rat

    Alcohol intoxication progressively impairs drivers' capacity to detect important environmental stimuli

    Get PDF
    Rationale Alcohol intoxication impairs driving skills, leading to an increased frequency of accidents and crash fatalities. Inebriation may specifically impair environmental vigilance, reducing the driver's capacity for attention to stimuli that are relevant to successful navigation. Objectives We examined the separate and interactive effects of breath alcohol concentration (BrAC) and simulated driving scenario on the capacity to correctly identify visual stimuli embedded in the environment. Methods Ten healthy young adult drivers (6 males; 4 females) each performed 4 driving scenarios at each of 3 steady breath alcohol concentration levels (0, 60 and 100 mg/dl). Scenarios were based on speed or distance keeping while navigating a rural 2-lane road in daytime or nighttime conditions. Drivers pressed a button on the steering wheel corresponding to the direction of an arrow (up or down) which appeared briefly on road signs embedded in the environment, either overhead or on the roadside. Results Increasing level of BrAC and subjective scenario difficulty manifested significant, separate, but not interactive influences in association with the number of arrows correctly identified. Significant impairments could be detected at a level of BrAC below the current American limit for legal operation of a motor vehicle. Conclusions Environmental vigilance is subject to impairment by either/both alcohol intoxication and driving conditions

    Drugs of Abuse Can Entrain Circadian Rhythms

    Get PDF
    Circadian rhythms prepare organisms for predictable events during the Earth's 24-h day. These rhythms are entrained by a variety of stimuli. Light is the most ubiquitous and best known zeitgeber, but a number of others have been identified, including food, social cues, locomotor activity, and, most recently drugs of abuse. Given the diversity of zeitgebers, it is probably not surprising that genes capable of clock functions are located throughout almost all organs and tissues. Recent evidence suggests that drugs of abuse can directly entrain some circadian rhythms. We have report here that entrainment by drugs of abuse is independent of the suprachiasmatic nucleus and the light/dark cycle, is not dependent on direct locomotor stimulation, and is shared by a variety of classes of drugs of abuse. We suggest that drug-entrained rhythms reflect variations in underlying neurophysiological states. This could be the basis for known daily variations in drug metabolism, tolerance, and sensitivity to drug reward. These rhythms could also take the form of daily periods of increased motivation to seek and take drugs, and thus contribute to abuse, addiction and relapse

    Circadian entrainment by food and drugs of abuse

    Get PDF
    Circadian rhythms organize behavior and physiological processes to be appropriate to the predictable cycle of daily events. These rhythms are entrained by stimuli that provide time of day cues (zeitgebers), such as light, which regulates the sleep-wake cycle and associated rhythms. But other events, including meals, social cues, and bouts of locomotor activity, can act as zeitgebers. Recent evidence shows that most organs and tissues contain cells that are capable of some degree of independent circadian cycling, suggesting the circadian system is more broadly and diffusely distributed. Within laboratory studies of behavior, circadian rhythms tend to be treated as a complication to be minimized, but they offer a useful model of predictable shifts in behavioral tendencies. In the present review, we summarize the evidence that formed the basis for a hypothesis that drugs of abuse can entrain circadian rhythms and describe the outcome of a series of experiments designed to test that hypothesis. We propose that such drug-entrained rhythms may contribute to demonstrated daily variations in drug metabolism, tolerance, and sensitivity to drug reward. Of particular importance, these rhythms may be evoked by a single episode of drug taking, strengthen with repeated episodes, and reemerge after long periods of abstinence, thereby contributing to drug abuse, addiction, and relapse

    Adaptation of Subjective Responses to Alcohol is Affected by an Interaction of GABRA2 Genotype and Recent Drinking

    Get PDF
    BACKGROUND: Subjective perceptions of alcohol intoxication are associated with altered risk for alcohol abuse and dependence. Acute adaptation of these perceptions may influence such risk and may involve genes associated with pleasant perceptions or the relief of anxiety. This study assessed the effect of variation in the GABAA receptor genes GABRG1 and GABRA2 and recent drinking history on the acute adaptation of subjective responses to alcohol. METHODS: One hundred and thirty-two nondependent moderate to heavy drinkers, aged 21 to 27, participated in 2 single-blind, counterbalanced sessions, approximately 1 week apart. One session was an intravenous alcohol "clamp," during which breath alcohol concentration was held steady at 60 mg/dl (60 mg%) for 3 hours, and the other an identical session using saline infusion. Subjective perceptions of Intoxication, Enjoyment, Stimulation, Relaxation, Anxiety, Tiredness, and Estimated Number of Drinks were acquired before (baseline), and during the first and final 45 minutes of the clamp. A placebo-adjusted index of the subject's acute adaptation to alcohol was calculated for each of the 7 subjective measures and used in a principal component analysis to create a single aggregate estimate for each subject's adaptive response to alcohol. Analysis of covariance tested whether GABRA2 and GABRG1 single nucleotide polymorphism (SNP) genotypes, gender, placebo session, family history of alcoholism, recent drinking history, and the genotype × recent drinking history interaction significantly predicted the adaptive response. RESULTS: Recent drinking history (p = 0.01), and recent drinking history × genotype interaction (p = 0.01) were significantly associated with acute adaptation of the subjective responses to alcohol for the GABRA2 SNP rs279858. CONCLUSIONS: Higher recent drinking was found to be associated with reduced acute tolerance to positive, stimulating effects of alcohol in carriers of the rs279858 risk allele. We postulate that the GABRA2 effect on alcohol dependence may, in part, be due to its effect on subjective responses to alcohol

    Translating Preclinical Models of Alcohol Seeking and Consumption into the Human Laboratory using Intravenous Alcohol Self-Administration Paradigms

    Get PDF
    Preclinical models of Alcohol Use Disorder (AUD) have advanced theoretical, mechanistic, and pharmacological study of the human condition. “Liking” and “wanting” behaviors reflect core processes underlying several models of AUD. However, the development and application of translational models of these preclinical approaches are at an incipient stage. The goal of this study was to examine how intravenous free-access and progressive-ratio, operant-response human alcohol self-administration paradigms can be used as translational human model parallel of preclinical “liking” and “wanting”. Participants were 40 adults (Mean age=23.7, SD=2.0; 45% Female) of European descent who reported 12.6 drinking days (SD=5.2) out of the previous 30 (average= 4.1 drinks/drinking day (SD=1.7)). Individuals diverged in their alcohol self-administration behavior, such that free-access and progressive-ratio paradigm outcomes were not significantly correlated (p=.44). Free-access alcohol seeking was related to enjoying alcohol (p.70) to preferred level of breath alcohol concentration (BrAC) in the free-access session, a measure of liking alcohol. Family history of alcoholism, disinhibition traits, and recent drinking history were significantly related (p’s<.05) to alcohol seeking in the progressive-ratio paradigm, a measure of wanting alcohol. We conclude that intravenous alcohol self-administration paradigms show promise in modeling behaviors that characterize and parallel alcohol “liking” and “wanting” in preclinical models. These paradigms provide a translational link between preclinical methods and clinical trials

    Binge and High-Intensity Drinking – Associations with Intravenous Alcohol Self-Administration and Underlying Risk Factors

    Get PDF
    Some styles of alcohol consumption are riskier than others. How the level and rate of alcohol exposure contribute to the increased risk of alcohol use disorder is unclear, but likely depends on the alcohol concentration time course. We hypothesized that the brain is sensitive to the alcohol concentration rate of change and that people at greater risk would self-administer faster. We developed a novel intravenous alcohol self-administration paradigm to allow participants direct and reproducible control over how quickly their breath alcohol concentration changes. We used drinking intensity and the density of biological family history of alcohol dependence as proxies for risk. Thirty-five alcohol drinking participants aged 21-28 years provided analytical data from a single, intravenous alcohol self-administration session using our computer-assisted alcohol infusion system rate control paradigm. A shorter time to reach 80 mg/dl was associated with increasing multiples of the binge drinking definition (p = 0.004), which was in turn related to higher density of family history of alcoholism (FHD, p = 0.04). Rate-dependent changes in subjective response (intoxication and stimulation) were also associated with FHD (each p = 0.001). Subsequently, given the limited sample size and FHD range, associations between multiples of the binge drinking definition and FHD were replicated and extended in analyses of the Collaborative Study on the Genetics of Alcoholism database. The rate control paradigm models binge and high-intensity drinking in the laboratory and provides a novel way to examine the relationship between the pharmacokinetics and pharmacodynamics of alcohol and potentially the risk for the development of alcohol use disorders

    The Influence of The Burrowing Mammal Tamias Striatus on The Soil Creep Process

    No full text
    160 p.Thesis (Ph.D.)--University of Illinois at Urbana-Champaign, 1985.The eastern chipmunk, Tamias striatus, was used as a representative burrowing animal to investigate the effect which organisms have on geomorphic slope processes in east central Illinois, U.S.A. Soil creep measurements were made on both a short-term (biweekly) and a long-term (one year) basis, using t-bars and Young pits respectively. Surface sediment transport was monitored with the use of sediment traps. Data on soil moisture, soil temperature, and precipitation were also collected. Mean soil creep rates for the 0.0 to 10.0 centimeter depth interval were calculated at 0.0 mm/yr (+OR-) 0.00 for the 0-5 and 10-15 degree slopes, 0.23 mm/yr (+OR-) 0.17 for 20-25 degree slopes, and 0.75 mm/yr (+OR-) 0.00 for 30-35 degree slopes. Creep rates for individual depth increments of 0.0, 5.0, and 10.0 centimeters were also calculated for the individual slope catagories. T. striatus was not found to have a statistically significant effect on soil creep rates over the one year study period. Downslope displacement of surface sediment was found to be intensified on 10-15 and 20-25 degree slopes when T. striatus was present. Slope angle, vegetation, precipitation, and soil character were found to have an effect on the amount of displaced sediment. A lack of defined freeze-thaw cycles and a long dry period precluded the drawing of significant conclusions concerning their relative importance to the soil creep process in the research area.U of I OnlyRestricted to the U of I community idenfinitely during batch ingest of legacy ETD
    corecore