text
stringlengths
25
68.8k
quality
float64
1
2.58
source
stringclasses
5 values
- 1 How much difference does a wetsuit make in triathlon? - 2 Is a wetsuit necessary for triathlon? - 3 Do triathlon wetsuits help you float? - 4 What thickness wetsuit do I need for triathlon? - 5 Is it faster to swim in a wetsuit? - 6 Is a tri suit a wet suit? - 7 Can you swim in a triathlon suit? - 8 Do you change clothes in a triathlon? - 9 What is the difference between a triathlon wetsuit and a normal wetsuit? - 10 Should I wear anything under a wetsuit? - 11 Can you drown in a wetsuit? - 12 What is warmer a wetsuit or drysuit? - 13 What water temperature do you need a wetsuit triathlon? - 14 How thick should a swimming wetsuit be? How much difference does a wetsuit make in triathlon? A slower swimmer doing an Iron distance can be in the water 2 hours, and want the warmth. That said, most triathletes want the wetsuit for the buoyancy, which makes you a faster swimmer, rather than the warmth. The slower you swim, the more the wet suit helps you. Is a wetsuit necessary for triathlon? In a triathlon you want to wear what is called a triathlon suit or a triathlon kit (a top/bottom combo) designed to be worn in the swim, the bike, and finally the run. In other words, you do not take it off! With the exception of the wetsuit (if needed), you need not put on any other piece of clothing. Do triathlon wetsuits help you float? Triathlon. People who compete in triathlons will appreciate the extra lift a wetsuit provides. Whatever activity you use a wetsuit for, they'll help you float – and keep you warm while doing it. What thickness wetsuit do I need for triathlon? USAT, ITU and WTC (Ironman) have rules restricting the thickness of the wetsuit to no more than 5mm. Most brands will put the thickest rubber on the front torso and legs to give the heaviest areas of your body the most buoyancy. Thinner rubber (1.5-2mm) at the shoulders allows for greater flexibility and arm speed. Is it faster to swim in a wetsuit? Wetsuits were invented to keep you warm – but they have another positive side effect: they make you faster, too. Tests with and without wetsuits have shown a time saving of five to ten seconds per hundred meters, with wetsuit. Simply put, a full-sleeve suit will help any swimmer go faster. Is a tri suit a wet suit? Wetsuit rules The wetsuit will help you keep warm in the colder water, hence on warmer swims the use of a wetsuit may be prohibited – in which case, you just swim in the tri suit. So, in the UK the vast majority of open water swims and triathlon events will allow wetsuit use. Can you swim in a triathlon suit? A "tri suit" is one choice of garment you can wear through the entire race. This helps make the suit more comfortable for swimming. After the swim, the thin pad dries out quickly and doesn't feel like a big, wet diaper on the bike ride. Do you change clothes in a triathlon? Clothing for a Triathlon You can opt to change into clothes specific to each leg of the race —or not. Some triathletes choose to do the whole race in a swimsuit for quicker transitions. Other swimmers simply pull on a pair of shorts before jumping on their bike. What is the difference between a triathlon wetsuit and a normal wetsuit? Wetsuits used for water sports are usually covered by abrasion-resistant Lycra and are not particularly designed for swimming positions. Triathlon wetsuits are designed for flexible swimming positions, making you comfortable especially when your arms are above your head and are pointed forward. Should I wear anything under a wetsuit? Wearing nothing under your wetsuit is totally acceptable, and a matter of personal preference. However, consider these factors: Chafing: Wetsuits can chafe. Give it a try, and wear a swimsuit if it's uncomfortable. Can you drown in a wetsuit? Yes. Surfers, scuba divers, kayakers, etc. have all drowned while wearing wetsuits. It'd be very difficult to sink while wearing a wetsuit, so the potential to drown is much lower. Just beware that the wetsuit may make you feel constricted which can lead to panic. What is warmer a wetsuit or drysuit? Wetsuits are made of rubber neoprene and are designed to keep you warm when wet, but unlike drysuits, they are not waterproof. So, if you have a loose fitting wetsuit you will get cold. Drysuits, on the other hand, are completely waterproof, but not designed for warmth if used alone. What water temperature do you need a wetsuit triathlon? USAT Wetsuit Rules USAT rules state that competitors may wear wetsuits if the water temperature is 78 degrees or lower. USAT rules also state that if the water temperature is between 78.1 – 83.9 degrees, competitors may wear wetsuits but will not be eligible for awards. How thick should a swimming wetsuit be? The rules state a wetsuit is made of material providing thermal insulation, shall completely cover torso, back, shoulders and knees. They shall not extend past neck, wrists and ankles. They should be between 3mm and 5mm thickness but can be lower in some parts of the suit to allow free movement.
1.063143
m-a-p/FineFineWeb
An Experienced, Established Legal Team – The Support You Need 1. Home 2. » 3. Family Law 4. » Dividing Liabilities Dividing Liabilities Believe it or not, there is no specific statute in Kentucky which tells the court how to divide marital debts. Generally, the debt "follows" the asset. But all debts incurred during a marriage are not automatically deemed to be "marital" in nature. Courts in Kentucky have a five-part evaluation in ruling whether a debt is marital or not: 1. Who benefited from the incurring of the debt? If it was both parties or the family, then the debt will likely be determined to be marital. 2. Who participated in incurring the debt? If one party took out a loan without the knowledge of the other and only the first party benefited from the loan, then the Court will probably determine that debt to be non-marital. The same would apply to running up a secret credit card balance, gambling debt, etc. 3. What was the purpose of the debt? Was it incurred to acquire something for the benefit of both parties or the whole family? If the debt was incurred to make improvements to the marital home, for instance, then the debt would likely be considered marital as well. If it was something that only benefitted one party (clothing, vacations, luxury items, etc.), maybe not. 4. Was the debt necessary to support the family? Maybe the debt was incurred purely by one spouse but the money was used to pay a child's medical expenses. That would probably be considered a marital debt. But maybe the debt was the result of one spouse's decision to acquire an asset, over the objection of the other spouse. For instance, if one of them decided to buy a boat or a sports car and the other party objected and did not benefit from the purchase, then that debt would presumably be determined to be non-marital in nature. 5. Which party has the greater financial resources? If one party has more assets or more income than the other spouse, then he or she has a much higher level of exposure in being held responsible for debt incurred during the marriage. Regardless of the above five factors, Kentucky courts still have the right to make decisions on the division of liabilities under their discretion as none of the above are considered "determinative". The answer to a single question about the debt does not create a hard and fast rule that the Court must obey. That's why having a Family Law attorney who has dealt with these complexities is so important to the equitable resolution of your divorce case. Generally speaking, however, the debts will end up with the party that gets the attendant asset. If you got the car or the residence, you'll probably end up with the debt as well. Most debts are divided as equally as possible. But, as you would expect, the spouse with the greater income often ends up with greater debt. A pendente lite ("temporary") action to limit the accumulation of new credit and debts should be seriously considered until the divorce case is completed. In some cases, it may be appropriate for your lawyer to immediately notify creditors about your divorce proceedings. It may also be appropriate to close certain credit card accounts to prevent new debt from been accumulated. Every divorce should culminate in the reaching of a resolution of each party's claims and an equitable distribution of the parties' joint assets as well as division of their responsibilities insofar as any children and/or debts are concerned. If you have questions about how to divide any debts you share or hold separately in your name and are contemplating a divorce, please contact our office.
1.25165
Zyphra/Zyda-2
Type of Document Dissertation Author Hagen, Thomas Ch. URN etd-110298-102457 Title Elongational Flows in Polymer Processing Degree PhD Department Mathematics Advisory Committee Advisor Name Title Renardy, Michael J. Committee Chair Baird, Donald G. Committee Member Herdman, Terry L. Committee Member Renardy, Yuriko Y. Committee Member Rogers, Robert C. Committee Member Keywords - Fiber Spinning - Linear Stability - Quasilinear Hyperbolic Equations - Spectral Determinacy Date of Defense 1998-12-01 Availability unrestricted Abstract The production of long, thin polymeric fibers is a main objective of the textile industry. Melt-spinning is a particularly simple and effective technique. In this work, we shall discuss the equations of melt-spinning in viscous and viscoelastic flow. These quasilinear hyperbolic equations model the uniaxial extension of a fluid thread before its solidification. We will address the following topics: first we shall prove existence, uniqueness, and regularity of solutions. Our solution strategy will be developed in detail for the viscous case. For non-Newtonian and isothermal flows, we shall outline the general ideas. Our solution technique consists of energy estimates and fixed-point arguments in appropriate Banach spaces. The existence result for a simple transport equation is the key to understanding the quasilinear case. The second issue of this exposition will be the stability of the unforced frost line formation. We will give a rigorous justification that, in the viscous regime, the linearized equations obey the ''Principle of Linear Stability''. As a consequence, we are allowed to relate the stability of the associated strongly continuous semigroup to the numerical resolution of the spectrum of its generator. By using a spectral collocation method, we shall derive numerical results on the eigenvalue distribution, thereby confirming prior results on the stability of the steady-state solution. Filename Size Approximate Download Time (Hours:Minutes:Seconds) 28.8 Modem 56K Modem ISDN (64 Kb) ISDN (128 Kb) Higher-speed Access hagenetd.pdf 687.09 Kb 00:03:10 00:01:38 00:01:25 00:00:42 00:00:03 If you have questions or technical problems, please Contact DLA.
1.196809
openbmb/Ultra-FineWeb
drug excretion is highly dependent on the glomerular filtration rate (GFR) and blood flow to the kidney. Since GFR is relatively constant among individuals with normal renal function, the elimination of drugs that are primarily excreted unchanged in the urine is also less variable. First-Order Elimination The rate constant of elimination (k) is the sum of the first-order rate constant for metabolism (km) and the first-order rate constant for excretion (ke): k = ke + km (12.1) 309 310 Chapter 12 »» Define Michaelis–Menton kinetics and capacity-mediated metabolism. »» Calculate drug and metabolite concentrations for drugs that undergo both hepatic and biliary elimination. »» Define first-pass metabolism and describe the relationship between first-pass metabolism and oral drug bioavailability. »» Use urine data to calculate fraction of drug excreted and metabolized. »» Explain how Michaelis–Menton kinetics can be used to determine the mechanism of enzyme inhibition and transporter inhibition. »» Describe biliary drug excretion and define enterohepatic drug elimination. »» Discuss the reasons why bioavailability is variable and can be less than 100%. »» Describe BDDCS—Biological Drug Disposition Classification System. In practice, the excretion rate constant (ke) is easily evaluated for drugs that are primarily renally excreted. Nonrenal drug elimination is usually assumed to be due for the most part to hepatic metabolism, though metabolism or degradation can occur in any organ or tissue that contains metabolic enzymes or is in a degradative condition. Therefore, the rate constant for metabolism (km) is difficult to measure directly and is usually obtained from the difference between k and ke. km = k - ke A drug may be biotransformed to several metabolites (metabolite A, metabolite B, metabolite C, etc); thus, the metabolism rate constant (km) is the sum of the rate constants for the formation of each metabolite: k m = kmA + kmB + kmC +  + kmI (12.2) The relationship in this equation assumes that the process of metabolism is first order and that the substrate (drug) concentration is very low. Drug concentrations at therapeutic plasma levels for most drugs are much lower than the Michaelis–Menten constant, KM, and do not saturate the enzymes involved in metabolism. Nonlinear Michaelis–Menten kinetics must be used when drug concentrations saturate metabolic enzymes (see also Chapter 21). Because these rates of elimination at low drug concentration are considered first-order processes, the percentage of total drug metabolized may be obtained by the following expression: % drug metabolized = km × 100 (12.3) k Fraction of Drug Excreted Unchanged (fe) and Fraction of Drug Metabolized (1 - fe) For most drugs, the fraction of dose eliminated unchanged (fe) and the fraction of dose eliminated as metabolites can be determined. For example, consider a drug that has two major metabolites and is also eliminated by renal excretion (Fig. 12-1). Assume that 100 mg of the drug was given to a patient and the drug was completely absorbed (bioavailability factor F = 1). A complete (cumulative) urine collection was obtained, and the quantities in parentheses in Fig. 12-1 indicate the amounts of each metabolite and unchanged drug that were recovered. The overall elimination half-life (t1/2) for this drug was 2.0 hours (k = 0.347 h-1). Drug Elimination and Hepatic Clearance kmA Drug (100 mg) kmB ke Metabolite A (10 mg) To determine the renal excretion rate constant, the following relationship is used: ke total dose excreted in urine Du∞ = = total dose absorbed k FD0 (12.4) where Du∞ is the total amount of unchanged drug recovered in the urine. In this example, ke is found by proper substitution into Equation 12.4: ke = (0.347) ke 0.243 × 100 = × 100 = 70% 0.347 k Alternatively, because 70 mg of unchanged drug was recovered from a total dose of 100 mg, the percent of drug excretion may be obtained by % drug excretion = 70 × 100 = 70% 100 Therefore, the percent of drug metabolized is 100%-70%, or 30%. For many drugs, the literature has approximate values for the fraction of drug (fe) excreted unchanged in the urine. In this example, the value of ke may be estimated from the literature values for the elimination half-life of the drug and fe. Assuming that the (12.5) ke = (0.7) (0.347) = 0.243 h-1 PRACTICAL FOCUS The percentages of drug excreted and metabolised are clinically useful information. If the renal excretion pathway becomes impaired, as in certain kidney disorders, then less drug will be excreted renally and hepatic metabolism may become the primary drug elimination route. The reverse is true if liver function declines. For example, if in the above situation renal excretion becomes totally impaired (ke ≈ 0), the elimination t1/2 can be determined as follows: k = km + ke but 70 = 0.243 h −1 100 To find the percent of drug eliminated by renal excretion, the following approach may be used: % drug excretion = ke = fek Because t1/2 is 2 hours, k is 0.693/2 h = 0.347 h-1, and ke is Unchanged drug in urine (70 mg) FIGURE 12-1 Model of a drug that has two major metabolites and is also eliminated by renal excretion. elimination half-life of the drug is 2 hours and fe is 0.7, then ke is estimated by Equation 12.5. Metabolite B (20 mg) 311 ke ≈ 0 Therefore, k ≈ km ≈ 0.104 h-1 The new t1/2 (assuming complete renal shutdown) is t1/ 2 = 0.693 = 6.7 h 0.104 In this example, renal impairment caused the drug elimination t1/2 to be prolonged from 2 to 6.7 hours. Clinically, the dosage of this drug must be lowered to prevent the accumulation of toxic drug levels. Methods for adjusting the dose for renal impairment are discussed in Chapter 24. HEPATIC CLEARANCE The clearance concept may be applied to any organ and is used as a measure of drug elimination by the organ (see Chapter 7). Hepatic clearance may be 312 Chapter 12 defined as the volume of blood that perfuses the liver which is cleared of drug per unit of time. As discussed in Chapter 7, total body clearance is composed of all the clearances in the body: ClT = Clnr + Clr (12.6) where ClT is total body clearance, Clnr is nonrenal clearance (often equated with hepatic clearance, Clh), and Clr is renal clearance. Hepatic clearance (Clh) is also equal to total body clearance (ClT) minus renal clearance (ClR) assuming no other organ metabolism, as shown by rearranging Equation 12.6 to Clh = ClT - ClR (12.6a) EXAMPLES • • • 1. The total body clearance for a drug is 15 mL/ min/kg. Renal clearance accounts for 10 mL/ min/kg. What is the hepatic clearance for the drug? Solution Hepatic clearance = 15 – 10 = 5 mL/min/kg Sometimes the renal clearance is not known, in which case hepatic clearance and renal clearance may be calculated from the percent of intact drug recovered in the urine. 2. The total body clearance of a drug is 10 mL/ min/kg. The renal clearance is not known. From a urinary drug excretion study, 60% of the drug is recovered intact and 40% is recovered as metabolites. What is the hepatic clearance for the drug, assuming that metabolism occurs in the liver? Solution Hepatic clearance = total body clearance × (1 – fe) (12.7) where fe = fraction of intact drug recovered in the urine. Hepatic clearance = 10 × (1 – 0.6) = 4 mL/min/kg In this example, the metabolites are recovered completely and hepatic clearance may be calculated as total body clearance times the percent of dose recovered as metabolites. Often, the metabolites are not completely recovered, thus precluding the accuracy of this approach. In this case, hepatic clearance is estimated as the difference between body clearance and renal clearance. Extrahepatic Metabolism A few drugs (eg, nitroglycerin) are metabolized extensively outside the liver. This is known as extrahepatic metabolism. A simple way to assess extrahepatic metabolism is to calculate hepatic (metabolic) and renal clearance of the drug and compare these clearances to total body clearance. EXAMPLES • • • 1. Morphine clearance, ClT, for a 75-kg male patient is 1800 mL/min. After an oral dose, 4% of the drug is excreted unchanged in the urine (fe = 0.04). The fraction of drug absorbed after an oral dose of morphine sulfate is 24%
1.744951
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
ованы в виде прикладной программы (или прикладных программ), хранящейся на считываемых посредством компьютера средах, таких как, например, постоянное запоминающее устройство на компакт-диске (CD-ROM), постоянное запоминающее устройство на универсальном цифровом диске (DVD-ROM), модуль флеш-памяти, или на иных материальных средах для хранения информации. Как правило, программы, выполняемые для реализации вариантов осуществления настоящего изобретения, могут являться частью операционной системы или конкретного приложения, компонента, программы, модуля, объекта или последовательности команд. Компьютерная программа из настоящего изобретения обычно состоит из множества команд, которые собственный компьютер транслирует в машинно-считываемый формат и, следовательно, в исполняемые команды. К тому же программы состоят из переменных и структур данных, которые либо хранятся локально относительно программы, либо находятся в памяти или в запоминающих устройствах. Кроме того, различные описанные ниже программы могут быть идентифицированы на основании того приложения, для которого они реализованы в конкретном варианте осуществления настоящего изобретения. Однако следует понимать, что любая приведенная ниже конкретная номенклатура программ использована просто для удобства, и, следовательно, настоящее изобретение не следует ограничивать его использованием исключительно в каком-либо конкретном приложении, которое определено и/или подразумевается такой номенклатурой. Как показано на чертеже, система 200 содержит прикладную программу 210 оптимизации, базу 215 данных о состоянии трубопровода, входные файлы 220 и выходные файлы 255. Входные файлы 220 могут содержать модель 225 системы трубопроводов, описывающую трубопровод, подлежащий оптимизации, ограничения, наложенные на трубопровод, цели оптимизации и параметры для процедуры оптимизации. Эта информация может также содержать сведения о текущем состоянии различных установок, установочные параметры компрессоров, сведения о текущей конфигурации многофункциональных элементов трубопровода и т.д. Как проиллюстрировано на чертеже, модель 225 системы трубопроводов содержит модель 235 сети трубопроводов и модели 230 установок. Модель 235 сети трубопроводов описывает топологию системы трубопроводов, подлежащей оптимизации, а модели 230 установок описывают одну или большее количество производственных установок, соединенных с системой трубопровод
1.222103
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
it designed for human beings at all? This is another hidden passage in Ellora caves that I tried to get through, but after 10 feet, it becomes so narrow that I can't go any further. Where do these mysterious tunnels lead? Who could have used such narrow passages? The other important question is: how can you carve such narrow passages if human beings cannot even get through them? Was it carved by humans at all? Were these carved for extraterrestrials that are smaller than human beings? Notice how this entrance is clearly going underground, in this cave temple. These are underground areas hiding in plain sight. The security guards told me that there are several underground tunnels in Ellora caves which gradually become too narrow for human beings to access them, and all of them are locked. From these old doors, we can assume that these tunnels have been closed from public access at least 30 to 40 years ago. These underground tunnels are not found in the same area, but are scattered over many different caves in Ellora, which stretches across a few miles. Is it possible that there could be a vast underground city like Derinkuyu in Turkey? Derinkuyu is an enormous underground city with advanced technical features that could accommodate more than 20,000 people, which was hidden for centuries only to be discovered in 1965. If it is true that there is a vast underground city in Ellora, shouldn't there be ventilation shafts that would allow air from the ground to reach below. In Derinkuyu, there are thousands of holes drilled from the ground that act as ventilation shafts to the underground city. Look at this long ventilation shaft in Ellora caves that is drilled secretly in a dark chamber. It is about 4 inches wide, but it is so deep that we can't even see the other end. Could this be a ventilation shaft that reaches the underground city? How about this ventilation shaft, which is also going underground? You can also see hundreds of holes like this drilled on the rock floor. Some of them are unfinished, and only few inches deep, but notice how the other holes have been closed with concrete recently. I asked why some of these holes have been closed and my tour guide told me that someone dropped their car keys into one of them, but these holes are so deep that they couldn't retrieve it. So, they just covered these holes by putting concrete over them. What else could be the purpose of these deep holes on the floor, if they are not ventilation shafts? Why would hundreds of holes be drilled on these rock floors with painstaking efforts without any reason? Now, this is one of the abandoned, bat infested places and let's take a look at what's inside. This seems like a very old sanctum holding a lingam, which is a Hindu god. Centuries ago, plenty of water would be poured on this lingam everyday as part of religious worship, and it would drain through this channel. Notice that this water drain has been closed recently with some rocks. Where would the water go? Let's see...It would go vertically down through this tunnel. Unless there is an area underground, this would be pointless because the water would just overflow all over the place! All over Ellora caves, there are hundreds of idols which would require thousands of gallons of water poured on them every day. Was this a technique to get clean water underground that could be used for different purposes?" Breaking News!! UFO Sightings Massive Alien Ship Visits ISS Caught On Video! Massive Alien Ship Visits ISS Caught On Video! Astonishing Footage and Incredible Detail! Original Nasa Footage Sunday, March 29, 2015 Alien Structure Found On Mars UFO News: UFO And Alien Structure Found In Mars Photo. -UFO AlienSource photo: Watch also: UFO flying near Pyramids on Mars BIGGEST Snake EVER (195 Ft.) Analyzed & Confirmed The Biggest Snake in the World recorded, seen, or filmed is analyzed in depth and confirmed to be around 195 feet in length and could easily eat a man. A monstrous snake nearly the size of a football field is captured on film as it rears back to attack a helicopter flying by, but could anything that big really exist? Follow my in depth analysis and by the end of this video try asking yourself that question again. Humanity is familiar with some pretty big snakes, both in person and in the fossil records, and sometimes through photographic evidence by reliable witnesses. During World War Two, Belgian fighter ace and squadron leader Remy Van Lierd rose to the rank of Colonel in the British Royal Air Force, and received the Distinguished Flying Cross medal of valor, after shooting down six enemy aircraft and forty four V-1 Flying Bombs. After the war he was put in command of the Kamina Air Base, in the Katanga region of what was then The Belgian Congo. His reputation and spatial recognition abilities are impeccable, which means we must give careful consideration to this incredible report. One day in Africa in 1959, on the way back to the airbase from a helicopter mission, Colonel Lierd was astounded to see a massive snake winding it's way along the terrain below. He circled around for a better look and had one of his flight crew take a series of photographs. As he descended to about 500 feet altitude, the gigantic snake took notice, and reared back as though ready to strike at the helicopter if it dared to come within range. By his own estimate, Colonel Lierd guesses that the snake must have been a minimum of 50 feet in length, with a triangular shaped jaw about two to three feet wide. In his own words, "It could have easily eaten up a man!" Keep in mind that a reticulated python is the longest snake ever measured alive at 33 feet or 10 meters in length. T he longest measured known extinct snake is the Titanaboa, which was 42 ft or about 13 meters in length. After "countless reports from the Kamina area of giant snakes", the original photos were sent to CIA Captain Lorenzo W. Burrows, director of the National Photographic Interpretation Center with some further analysis of the photo. Let's look at that. Here is the full photo in it's entirety. As you can see at the bottom, the type of camera the photo was taken with was recorded, which adds some authenticity to this photo. The key factor in this photo, as pointed out in the letter, are the termite nests, which is a little confusing at first unless you know what an African termite nest looks like. They look like this…they're absolutely huge and they take on recognizable shapes. The letter reads "Further, the two black spots in the upper left hand quadrant are quite plainly termite nests, or rather, their shadows. Under high power you can very easily see this, and that the top one is a column, circular in the horizontal section, while the bottom one is typical of the flange types that always point north to south." See the termite nests? Now remember how big they are? Let's use them, as is done in the letter, to approximate a scale of reference. These termite mounds can easily be 2 and a half times the size of a human. So, we'll say the top termite nest is 15 feet or about 4.5 meters tall. This is a bit of a guess, but if the snake was fully stretched out, I'd say it'd be around this long. So if we just stack our 15 foot or 4.5 meter termite nests next to this line…we get about 13 of them. This yields a rough approximate total length of about 195 feet or about 60 meters and closely matches what the letter says of about 200 feet. That's up to 5 times larger than even the Titanaboa…allegedly the biggest snake ever and that was thought to have gone extinct shortly after the dinosaurs. What sort of monster did Colonel Remy Van Lierd encounter and photograph here?...and why are there so many other alleged reports of massive snakes by others in the same area? The Congo seems like a very, very scary place. Crave more mysteries? Watch our Top 5 BIGGEST Animal Mysteries EVER video if you havn't already or jump on our mystery playlist and discover how aliens may have saved Earth from a nuclear catastrophe….twice.
1.449705
Zyphra/Zyda-2
Europeans are plagued by mental and neurological illnesses, with almost 165 million people or 38 percent of the population suffering each year from a brain disorder such as depression, anxiety, insomnia or dementia, according to a large new study. With only about a third of cases receiving the therapy or medication needed, mental illnesses cause a huge economic and social burden -- measured in the hundreds of billions of euros -- as sufferers become too unwell to work and personal relationships break down. "Mental disorders have become Europe's largest health challenge of the 21st century," the study's authors said. At the same time, some big drug companies are backing away from investment in research on how the brain works and affects behavior, putting the onus on governments and health charities to stump up funding for neuroscience. "The immense treatment gap... for mental disorders has to be closed," said Hans Ulrich Wittchen, director of the institute of clinical psychology and psychotherapy at Germany's Dresden University and the lead investigator on the European study. "Those few receiving treatment do so with considerable delays of an average of several years and rarely with the appropriate, state-of-the-art therapies." Wittchen led a three-year study covering 30 European countries -- the 27 European Union member states plus Switzerland, Iceland and Norway -- and a population of 514 million people. A direct comparison of the prevalence of mental illnesses in other parts of the world was not available because different studies adopt varying parameters. Wittchen's team looked at about 100 illnesses covering all major brain disorders from anxiety and depression to addiction to schizophrenia, as well as major neurological disorders including epilepsy, Parkinson's and multiple sclerosis. The results, published by the European College of Neuropsychopharmacology (ENCP) on Monday, show an "exceedingly high burden" of mental health disorders and brain illnesses, he told reporters at a briefing in London. Mental illnesses are a major cause of death, disability, and economic burden worldwide and the World Health Organization predicts that by 2020, depression will be the second leading contributor to the global burden of disease across all ages. Wittchen said that in Europe, that grim future had arrived early, with diseases of the brain already the single largest contributor to the EU's burden of ill health. The four most disabling conditions -- measured in terms of disability-adjusted life years or DALYs, a standard measure used to compare the impact of various diseases -- are depression, dementias such as Alzheimer's disease and vascular dementia, alcohol dependence and stroke. The last major European study of brain disorders, which was published in 2005 and covered a smaller population of about 301 million people, found 27 percent of the EU adult population was suffering from mental illnesses. Although the 2005 study cannot be compared directly with the latest finding -- the scope and population was different -- it found the cost burden of these and neurological disorders amounted to about 386 billion euros ($555 billion) a year at that time. Wittchen's team has yet to finalize the economic impact data from this latest work, but he said the costs would be "considerably more" than estimated in 2005. The researchers said it was crucial for health policy makers to recognize the enormous burden and devise ways to identify potential patients early -- possibly through screening -- and make treating them quickly a high priority. "Because mental disorders frequently start early in life, they have a strong malignant impact on later life," Wittchen said. "Only early targeted treatment in the young will effectively prevent the risk of increasingly largely proportions of severely ill...patients in the future." David Nutt, a neuropsychopharmacology expert at Imperial College London who was not involved in this study, agreed. "If you can get in early you may be able to change the trajectory of the illness so that it isn't inevitable that people go into disability," he said. "If we really want not to be left with this huge reservoir of mental and brain illness for the next few centuries, then we ought to be investing more now." (Reporting by Kate Kelland; Editing by Matthew Jones)
1.646329
HuggingFaceFW/fineweb-edu
one or two of the stanzas to Sorli. Lines 1-2 are quoted in the Volsungasaga. The manuscript does not indicate line I as beginning a stanza. Erp: Hamther means that while the two brothers had succeeded only in wounding Jormunrek, Erp, if he had been with them, would have killed him. Lines 3-4 may be a later interpolation. Norns: the fates; the word used in the original means the goddesses of ill fortune.] p. 555 Like the hounds of the Norns, | that nourished were In greed mid wastes so grim. 30. "We have greatly fought, | o'er the Goths do we stand By our blades laid low, | like eagles on branches; Great our fame though we die | today or tomorrow; None outlives the night | when the Norris have spoken." 31. Then Sorli beside | the gable sank, And Hamther fell | at the back of the house. This is called the old ballad of Hamther. [29. This is almost certainly an interpolated Ljothahattr stanza, though some editors have tried to expand it into the Fornyrthislag form. Hounds of the Norns: wolves. 30, Some editors assume a gap after this stanza. 31. Apparently a fragment of a stanza from the "old" Hamthesmol to which the annotator's concluding prose note refers. Some editors assume the loss of two lines after line 2. Prose. Regarding the "old" Hamthesmol, cf. Guthrunarhvot, introductory note.] Next: Pronouncing Index Of Proper Names
1.188854
Zyphra/Zyda-2
. Using dedicated vacuum hoses like TruBlu from Accutools will help consistently achieve deep vacuum quickly. Vacuum hoses are designed specifically to hold under deep vacuum and to keep moisture and contaminants from bonding to the inside. A typical refrigerant hose used for recovery and charging has been exposed to refrigerant, oil, moisture, and other contaminants. The micron gauge should be located as close to the system as possible for an accurate reading. When you place a micron gauge at the pump, it is measuring the pressure of the pump, which can be drastically different than the vacuum level at the system itself — especially at the furthest point in the system from where the pump is connected. We find that the best place to locate the vacuum (micron) gauge is either on the side port of the suction CRT when vacuum hoses are connected to both service ports, or directly on the liquid line port when using a single hose on the suction port. After degassing and dehydrating the system as deeply and as quickly as possible, isolate the system and micron gauge from the pump and hoses and confirm that it will hold the vacuum. This is easily done with CRTs by closing the valves after achieving the target vacuum level. Sometimes, air can get trapped in the ball valves of the CRTs and cause a significant rise when closing them. Slowly close and open the CRTs a couple of times before starting the decay test. In general, on a newly installed residential split system, you should be able to pull to a target vacuum of 300 microns or less. Then, close the CRT valves. The pressure in the system will rise slowly; this is called "vacuum decay." After at least 10 minutes of isolation, the system pressure should remain under 500 microns. When servicing an existing system, especially when the entire system, including the compressor, is being evacuated, a target vacuum level of 500 microns and decay to under 1000 microns after a minimum of 10 minutes of isolation is more realistic. If there are any leaks, trapped refrigerant, or moisture, then these targets will be impossible to achieve with typical residential or light commercial equipment up to 5 tons in capacity. Some manufacturers advise pulling down to certain levels, then breaking the vacuum with nitrogen to bring the system up to atmospheric pressure (0 psig/14.7 psia,) then repeating the process. This certainly isn't a bad practice, especially when contamination from moisture or interference from refrigerant is suspected. If moisture or other vapor contaminants are suspected, a "nitrogen sweep" can be performed between evacuations. This consists of bringing the system pressure up slowly from a vacuum to 0 psig/14.7 psia by injecting nitrogen into one port until the system is at or slightly above atmospheric pressure, then venting the other port to the atmosphere while continuing the nitrogen flow. This allows nitrogen to carry unwanted vapors out of the system by displacement. Some techs think that nitrogen "absorbs" moisture, but it does not. Nitrogen flowing through the system can carry moisture out or help to release refrigerant from the oil through entrainment, but it doesn't absorb anything. It does help to displace the air and the moisture in the air, which is worthwhile in some instances. You may hear that pulling a vacuum too quickly results in water freezing in the system. While it is true that water can freeze under vacuum, this only happens in real life when significant liquid water is in the system and when ambient temperatures are already near or below freezing. In low ambient conditions, it is advised to use a heat gun to warm the accumulator and evaporator and turn on the crankcase heater to help drive out the moisture. Pulling a deep and fast vacuum is always a good idea, and in a very wet system, using a heat gun and periodically sweeping with nitrogen will also help. Some techs say that pulling a vacuum below 250 microns can damage the compressor oil. I've done extensive research and spoken to many experts and have seen no evidence that evacuation below 250 microns causes any issues with POE or mineral oil. One excuse I often hear is that the ports on the system are ¼ inch, so larger hoses don't matter. This simply isn't true; larger, shorter hoses with no manifold can easily reduce vacuum time by 10 times or more, and this has been demonstrated time and time again. Even with a ¼-inch "choke point" in a few spots, the size of the hoses still matters significantly. - Use proper assembly practices to keep everything clean, dry, and tight. - Remove or fully depress cores. - Use dedicated large-diameter vacuum hoses and keep them as short as possible. - Keep clean oil in your pump and test the pump regularly. - Nix the manifold when evacuating and connect straight from the pump to CRTs. - Isolate and test after the desired vacuum level has been achieved to ensure there is no moisture or leaks. Using this method, you will find that newly installed residential equipment can be pulled down to under 500 microns in under five minutes with another 10 minutes for the decay test. This is a combination of time savings and best practices for a win-win result for you and the customer. Publication date: 2/25/2019
1.475438
Zyphra/Zyda-2
. In other words, this is a process in which society "pushes" and technology "pulls," with both processes embodied by an object whose technological makeup is relatively simple. This solution, which combines design and technology, was thus born of economic, social, and cultural changes during the period in question. The term "wearable technology" is often used today to describe the infiltration of cutting?edge technologies into the realm of objects worn on our bodies throughout the day. These include augmented reality eyewear, clothing made of smart textiles, bracelets and rings that monitor physiological activity and metabolism, shoes that change their appearance or communicate physiological information to a smartphone, and more. Yet to a large degree, eyeglasses were the first wearable technology, which may well be how they were perceived in the early Renaissance period (Veryat et al., 2008). With the exception of the frame, which has changed over time in accordance with new historical developments in the production of materials, lenses have remained the focal point of all eyewear. The technology of lens production made this product appear simultaneously threatening and fascinating during the early Renaissance, since it interfaces with the body and is carried on the face, causing us to almost forget its existence (until the glasses break or become dirty) – or, in other words, a wearable technology. Design became the central tool in shifting the cultural and functional associations of glasses from the world of medicine to the world of fashion and lifestyle. As technology and the use of materials continued to evolve, this shift became increasingly significant. If we take, for instance, contact lenses or modern frameless glasses, which represent attempts at assimilating physical signs of a "default," such attempts can be contrasted with the inverse phenomenon of underscoring the default by means of conspicuous design. Large glasses with massive black plastic frames have become a hipster accessory in recent years – a trend that can also be seen in the design of medical products. When coping with physical impairments, designers rely on one of two strategies – underscoring or blurring the defect. When they opt for blurring the impairment, designers attempt to assimilate the object into the overall appearance of the user's body, or to hide it under clothes. By contrast, when they choose to underscore it, designers use materials, textures or colors to transport the product from the world of medicine to the world of sports, extreme gear, or fashion (Ventura and Gunn, forthcoming). Indeed, ever since the 1940s, glasses have served as a fashion accessory associated with a certain socioeconomic status and level of intelligence (Segrave, 2011). Turkle speaks of "evocative objects" that have the power to create an effect of desire among consumers (2011). One may thus identify, in the course of the twentieth century, a growing tendency to purchase designer glasses that reflect the power of brands and further blur the origin of this complex product as a tool for corrective vision. Returning to the subject we started out with, which was the transition of eyeglasses from the realm of medicine to that of consumer culture, we can note the role played by designers in the transformation process of this common accessory. As an object that combines style, fashion, consumerism, and medicine, eyeglasses offer a fascinating case study of a double transformation: On the one hand, they have been transformed from a medical object into a consumer object. At the same time, their widespread use has led to their industrialized production out of cheap, reliable, durable materials (chiefly plastic), in contrast to the initial process of manual production out of metal or expensive materials such as tortoiseshell (Delong and Daly, 2013). Eyeglasses have thus become one of the objects that define our identity, and in many cases are even buried with us (Harper, 2012). Inversemedicalization: Underscoring a default by means of Design As already noted, since eyeglasses are almost the sole object that moved so dramatically from the realm of medicine to that of fashion, it is interesting to follow this trajectory through an examination of the body, and more specifically of the face. In contrast to evolutionary processes of assimilating and absorbing technologies into the body (similar to the process undergone, for instance, by car headlights, which have been assimilated into the body of the car to become part of its outline, rather than an external accessory hanging on its frame), the endurance of eyeglasses and the enhancement of their status is a unique phenomenon. Despite the successful development of a widely?used technology that enables the solution to be assimilated into the body (contact lenses), which seemingly "solved" the need for glasses to be perched on the user's nose, the use of eyeglasses has not diminished, and has in fact increased. This phenomenon becomes even more interesting when viewed in light of the developing trend to create technological products for medical purposes that are assimilated into our bodies, transforming us into cyborgs of sorts, to use the well?known term first employed by Donna Haraway (2006). In contrast to the bionic prosthesis, an artificial eye with technologically enhanced qualities and pacemakers, eyeglasses are here to stay, and we must remember that from a purely design?based perspective, their form has not changed for hundreds of years. Friedrich Herlin, Saint Peter reading with eyeglasses Altarpiece detail, 1466, St. Jacob's Church, Rothenburg ob der Tauber, Germany One motivation for this design?related nostalgia is obviously comfort, yet there are several additional reasons one can list as preserving and bolstering this phenomenon. To begin with, glasses provide the best excuse to don a mask. The moment we put them on, they do not only correct our vision and change the way we see the world, but also transform the way the world sees us. As already noted in the opening section of this article, glasses are to a large extent the first and most prominent wearable technology, which has endured for hundreds of years. Moreover, together with the makeup and jewelry we wear, glasses remain the most conspicuous design accessory seen by those who observe us, and play a central role in defining our taste and our professional and socioeconomic status. Secondly, there is no doubt that in addition to physically protecting the eyes, glasses also provide mental protection by mediating between our face, expressions and emotions and the outside world. In an urban environment that is increasingly charged with cognitive, visual, and auditory stimuli (as already discussed in Georg Simmel's classical essay on urban life), eyeglasses are a carefully designed life buoy. The use of glasses as a mask of sorts, even if not in the classical sense of the word, enables us to change our body image (recall the Duke of Milan, Francesco Sforza, who actively and self-reflexively collected dozens of pairs of glasses despite his normal vision). The same glasses may provide one person with a "serious" look, while endowing another one with a "naïve" appearance. Moreover, during certain periods glasses become an almost obligatory fashion accessory, as made evident by the now?hackneyed image of the black plastic glasses worn by Brooklyn hipsters. This legitimate mask, which functions as an easily replaceable accessory, enables us to change our preferred image of ourselves on an almost daily basis, in accordance with the event in question, our mood, and current trends. In this sense, then, eyewear design is not only a medical tool of the first order, but also a means of actively shaping our identity to fit changing situations throughout the day. People love changes and are constantly searching for novelty, even if only in terms of altering a frame's color or replacing a rounded frame with a pointed one. Whereas the medical, corrective component of eyeglasses is transparent and almost invisible, its supporting frame receives excessive and meticulous attention to its color, materiality, and form, whichhighlight the "importance" of the transparent circles it surrounds. The entire "fashion charge" of the glasses rests on the frame, so that that most eyeglass wearers spend more time choosing the frame than they do attending to the technological and optometric advances that produced the lenses, thus enabling them see the world. Although this cultural "carrier" is related to technological developments in the production of thin, light, and precise lenses, it also evolves independently to create variegated eyeglass forms and numerous new possibilities, much beyond the regular function of glasses for "corrective vision." This phenomenon itself is a unique case worthy of being noted. If we choose to read glasses as a signifier, focusing on the frame rather than on the lenses (the functional component), terms from the world of semiotics will inevitably enter the discussion. As we noted in a previous article (Ventura and Shvo, 2016), glasses – like every other design object – are composed of materials, colors and forms, which are all imbued with specific sociocultural meanings. We can thus state that the functional aspect of glasses (enabling people to see) is strong enough to allow for a minimal semiotic play of colors, forms, and materials. Regardless of whether they are made of horn, tortoiseshell, plastic, metal, or laminated wood, or printed using a 3D printer, eyeglasses are the most important and visible accessory we own, reflecting technological and perceptual changes related to social norms and changing fashions. Eyeglasses seem to be here to stay, even if we will need new excuses to leave this "prosthesis" perched on the tip of our noses. Jonathan Ventura, PhD, is a research fellow at the Helen Hamlyn Centre for Design, Royal College of Art, London. He is a senior lecturer in the Department of
1.996224
openbmb/Ultra-FineWeb
Physical Metallurgy and Mechanical Properties of Materials: Brittle Fractureby B. L. Averbach, Serial Information: Journal of the Engineering Mechanics Division, 1960, Vol. 86, Issue 6, Pg. 29-44 Document Type: Journal Paper At Symposium on Physical Metallurgy and Mechanical Properties of Materials, held by Eng Mechanics Div for summarizing current state of knowledge and some most recent developments in understanding of materials behavior, following paper covers Brittle Fracture. Subject Headings: Physical properties | Brittleness | Cracking Services: Buy this book/Buy this article Return to search
1.109579
openbmb/Ultra-FineWeb
The Unusual Explanation of How Rust Got on the Moon Vote for this video by social sharing! Hello and welcome! My name is Anton and in this video, we will talk about new discovery that suggests our planet has been "rusting" the surface of the Moon for billions of years. Tags: anton petrov, science, physics, astrophysics, astronomy, universe, whatdamath, what da math, technology, steven universe, space engine, moon, the moon, hematite, rust, the rust, rusting, rust on the moon, moon rust, earth is rusting the moon, oxygen from earth, iron oxide, levitating lunar dust, static on the moon, geocorona, earth corona, earth atmosphere and the moon, hemtatite on the moon, moon hematite, oxygen leaking from earth, oxygen escaping earth, earth magnetotail 1. Bob Thompson Iv said it before but I got to say it again. "Space out" is the coolest out tro. Period. Its the standard for all intros and outtros should be held to. 2. Josiffrank Must be all the Oxygen from the Moon landings all those Apollo rockets! Rusted our poor Moon rust forming in the moon is a sign that it was after all engineered. its mostly made of titanium. tell where the heck did it all come from? keep in mind that titanium to that extent of concentration has to be manufactured. 4. Donovan Piko The static electricity could also be part of why Apollo's comms had issues during landing, and could be potentially useful as an energy source and navigational tool. 5. Nick Diamond Does earth have a panspermia type effect on the moon? 6. Tumble Bug Love your channel, Anton! For me, Electric Universe Theory has the best explanation of moon phenomena, *period.* The moon's hematite being predominately on the side that is tidally locked with Earth should suggest to us that an electromagnetic phenomena is being observed. 7. Walter ZAMBOTTI So the oxygen from the Earth managed to reach all the way out to the moon but not so far as the other side of the moon. RRRRRIIIIIIIIGGGGGGGHHHHHHTTTTTTT!!!!!!! 8. esphi LEE That day when the American Submarine blown the japanese warship up into the moon…… that is why they never found the wreckage. 9. Efx245 Goodrich Remember we've already been to the Moon. 10. Blueblood1270 The moon is a bit rusty …. "Let's send some WD40 to lubricate the moon!!". 11. Julio C It's probably some failed chinese project I love the topics that you cover, and the amazing way you explain and cover them. Thank you so much! I love your channel. I know that there is more than 1 color of Hematite. I found that this was a fascinating topic. 13. anthony summers hmmm, static electricity, electro-megnetic shields, and plasmatic nuclear reactors for stars but electric universe isnt real and dark matter that has no evidence is real??? 14. jojones jojo Perhaps someone left it out in the rain. I'm looking at you Neil Armstrong. 15. James Green Or the moon is an artificial structure and now rusting :) 16. Denis Gantsev Rust getting on the moon. C++ shits bricks 17. XRP 747E Thank you as always, Anton! 18. janusz delondre this would mean that humans have never left the earth's atmosphere. 19. Rolla Coasta Ride the shadow of the moon is actually a perfect cone, and that's why wicked witches or wizards are portrayed wearing pointy hats, the space underneath the hat represents the mystery unknown shadow of the moon, occultism is light that is blocked forcing the dark shadow cone zone, hidden, obscured, secret, etc, the apollo astronauts all sensed a menacing presence when they were passing through the shadow of the moon, occultic vibes? 20. Contamination PD Those few spots on the dark side could be from particles getting snatched by the moons gravity and drawn to the surface 21. Rastrman TheSeer It's rusted out George Foreman grills. The aliens were trying chicken sliders there, back in the 1950's. 22. PowWowChicken Earth is a space ship and so is the moon. We are traveling with the mother ship called the sun 23. Per Nielsen Rust is in much desert dust – and is paramagnetic… it could come from the Earths deserts. 24. Duane Borgaes I'm sure the moon and all its glory exstends beyond our conventional understanding. 25. RSK Phosphine on Venus, Iron on the Moon, watch next they might find intelligence on Earth. 26. llessuR kraM noslekcim "Cosmic Rust!" Don't tell Megatron. 27. Fleem Q Swipes I'd guess it's more around the craters because that's where most of the surface iron is. 28. VMAXBX There is no dark side of the moon, it's all dark. 29. wozzapoet Loving the wonderful person cushion we bought for our baby, Anton, a beautiful and colourful welcome to the world to our little wonderful person, thanks mate🖖🏼 30. battlekruiser ok so there are clouds of charged particles near the lunar surface, right? inb4 lunar lightning 31. ViScAsh Would the Coanda effect of the solar wind or the magnetotail be responsible for depositing oxygen particles on the dark side? 32. TargaWheels They don't want us to see the far side of the moon. Because thats where Jimmy Hoffa is. 😄 I'm sure the nut cases have been saying this for years not the same reasoning of course, yours is actually way more far fetched 😂 34. Space Ghost So that's where my Jeep CJ-7 is slowly disappearing to 35. Bad-Bunny Blogger Back up Anton, the moon has static energy, can we stick a plug socket in it? 🤔 36. EeRocKK You're the wonderful person, really. You're just projecting every time you say that 😎😄🤟 37. Asmo It effects the craters more because that's where the iron is. From iron asteroids. 38. CounterLink I was early this time! 39. Aerrin Longheart I wonder if this increases the possibility of terrestrial life on the moon, and if so, what'll we think when we suddenly find it. 40. Dark Madder Smeone rubbed the moon on their head and then stuck it to the sky… 41. DragonsFire 3174 His smile at the end always gets me, I watch to the end just to see it, it shows he really loves what he does 42. arno kosterman Space dansety on the dark side is diverend en the diverentional bitwean de night en day is match bicker than the eurth side of the moon night en day is less diverend. Like last weak to less space damcety it metterstade becomes trajectory en becouse the moon feeds that wat is feeding itself whit it comes out where it feeds itself from. Olso the qwestion risses: our dit it materialised itself there on the spot, from the feilds of the iron in the eurth its exstarnal inursha🌐💟🌐 En the moons its own inusja🌞🍀💮 Our in the middle en dit the moon feeds it vor it to be reseaved👀💟 43. Billy I'm pretty sure it's because the moon is an an ancient space ship. 44. William James Rapp Thank you for another fantastic presentation. If I had money I would Pay you to keep on keepin on. 45. Gerard 2:11 the distance to the moon is three hundred and thousand kilometers PER SECOND ;-) 46. Panspermia Hunter The whole moon passes through the magnetotail so some will be caught by the far side of the Moon surely even if it is drag back to the Moon surface by the Moons gravity? that would be my thoughts on why rust is also found on the far side of the Moon. A fantastic discovery though! PS: why mostly on crater edges? probably because they are raised edges that will collect more oxygen than the smoother areas rather like swatting the slow or static oxygen as the Moon travels through the magnetotail, if you can imagine the edges are less aerodynamic because they catch the air (oxygen). 47. Sh
1.323166
Zyphra/Zyda-2
Ingenuity (Adjectives) Grammar — Advanced Level Share this exercise Complete each sentence choosing the correct option 1. Ingenuity is a robotic helicopter located on Mars since February 18, 2021. 2. The flights would get more as experience is gained and the allotted time for operating the helicopter dwindles. 3. The first flight is planned to be a hover at an altitude of 3 m, lasting about 40 seconds. 4. After deployment, the rover is expected to drive approximately 100 m away from the drone to allow it a "buffer zone." 5. For flight testing, a vacuum chamber was used to simulate the very low pressure of the atmosphere of Mars. Practice your writing skills by discussing the questions below 1. What is Ingenuity? 2. What are three interesting things that you want to discover in the outer space? Why? 3. If given a chance to travel to the outer space. where would you go and why? Need help? Ask a question or reserve a class with Jennifer From English No translation
2.102881
Zyphra/Zyda-2
improve the state's fishing and hunting opportunities and protect coastal communities. This report focuses on four major projects built or planned using Deepwater Horizon penalties that have directly benefited anglers and hunters by improving coastal habitats. All totaled, they represent an investment of more than $600 million in making Louisiana's "Sportsman's Paradise" a world-class hunting and fishing destination for decades to come. Caminada Headlands A.K.A. Elmer's Island The restoration of the Caminada Headlands is the largest coastal restoration project in Louisiana and the largest single investment in the recovery of the Gulf Coast after the 2010 Deepwater Horizon oil spill. Known by locals as Elmer's Island and Fourchon Beach, the Caminada Headlands is a 13-mile stretch of beaches, dunes, and marsh that extends from the mouth of Bayou Lafourche—called Belle Pass—all the way to Caminada Pass on the western tip of Grand Isle. This habitat was restored in two phases over five years using a combination of $145.9 million in oil spill fines from the National Fish and Wildlife Foundation's Gulf Environmental Benefit Fund plus $40 million from the Coastal Impact Assistance Program and $30 million in state budget surpluses. Without question, Caminada is Louisiana's most popular summertime fishing destination. Its proximity to Grand Isle—Louisiana's only inhabited barrier island—and Port Fourchon—one of the Gulf's most popular angling jumping-off spots—means hundreds of thousands of anglers, crabbers, and beach combers visit Elmer's and the Fourchon each year. The area is also home to Elmer's Island Wildlife Refuge, a public recreation area overseen by the Louisiana Department of Wildlife and Fisheries, which is popular among anglers and bird watchers. For the last century, hurricanes, strong winter storms, subsidence, and tidal currents have eaten away at the headland, causing the beach to retreat about 35 feet per year. These conditions threaten the more fragile marshes to its north, not to mention the energy infrastructure of Port Fourchon and camps and homes on Grand Isle and the only access road, Highway 1. In 2010, oil from the Deepwater Horizon spill coated this beach. Many of the iconic pictures circulated in the media coverage of the spill, showing sheets of sticky, rust-colored tar mats and brown pelicans coated with oil, were taken at Elmer's Island and the Fourchon Beach. Wading wasn't allowed for more than two years, as heavy equipment and cleaning crews scoured the sand for tar mats. Louisiana's Coastal Protection and Restoration Authority were back with more heavy equipment and personnel in August 2013, this time to restore the beaches and dunes, rather than drag and sift them for oil. The project used about 8.4 million cubic yards of sand mined from an ancient submerged delta of the Mississippi River to extend the beach 500 feet into the Gulf of Mexico, rebuild dunes up to seven feet tall, and plant those dunes with native vegetation. An additional 200 acres of tidal marsh were rebuilt, as well. Barges moved the sand from Ship Shoal to a staging area in Belle Pass where it was pumped onto the beach and then shaped with bulldozers. Sand fencing was placed along the beach to help capture wind-blown sand to rebuild dunes as well. By the summer of 2017, Elmer's was opened completely to wade fishing, restoring a summertime tradition that dates back several generations. Queen Bess Island Nearly $19 million in fines from the Natural Resources Damage Assessment have been invested in Queen Bess Island, located just two miles east of Grand Isle in Barataria Bay, to repair decades of erosion and subsidence, as well as oil spill damage. While far smaller in scope compared to the Caminada project, this restoration work has had outsize impacts for anglers and Louisiana's most iconic bird. After DDT and other pesticides decimated pelicans in the 1950s and early 1960s, Louisiana was left without a native population of the bird emblazoned on its state flag. In 1971, biologists from Louisiana and Florida worked together to bring 750 young Brown Pelicans to the Barataria Basin, releasing them on Queen Bess, where 11 pairs originally built their nests. By 2009, there were an estimated 80,000 or more Brown Pelicans in Louisiana and the bird was removed from the Endangered Species list. But only a year later, several thousand pelicans and critical habitat on Queen Bess Island were harmed by the Deepwater Horizon spill. The island is now home to the largest population of brown pelicans in the Barataria Basin, and nesting habitat has grown from just five acres in the summer of 2019 to 36 acres upon completion of the project in late February 2020. Queen Bess's proximity to Grand Isle also means that it is visited by tens of thousands of anglers every year. The Barataria Basin was once dotted with small islands like Queen Bess, the remnants of old headlands and bayou banks once connected to the Mississippi River that erosion and subsidence has wiped from the landscape. This has meant less land to dampen waves and protect fish and fishermen, making the island even more important to anglers. Construction for the Queen Bess project was compressed into a six-month period from late summer 2019 to late February 2020 to avoid interfering with pelican nesting season. It included ringing the island with limestone chunks up to three feet above sea level to give it additional capacity to withstand wave action, subsidence, and sea-level rise. Rock breakwaters were also built on the southwest shoreline to provide calm water areas for pelicans and other birds. Then, more than 150,000 cubic yards of sediment dredged from the Mississippi River was transported about 25 miles to the island on barges. After the sediment was shaped by bulldozers, tidal sloughs were constructed to allow some water flow into the island and native vegetation was planted to enhance habitat and fight erosion. The restoration project is expected to extend the life of the island for more than three decades, allowing brown pelicans to thrive where anglers can chase speckled trout, redfish, and sheepshead for years to come. Louisiana's Coastal Protection and Restoration Authority has invested nearly $500 million in rebuilding more than 60 miles of beaches and barrier islands since 2007. It has also overseen construction of hundreds of acres of back-barrier marshes designed to enhance fisheries habitat and help retain the sand that has been pumped ashore by dredges to rebuild beaches and dunes. Arguably, the most ambitious of these efforts, completed in 2018, is the restoration of the Caillou Lake Headlands in Terrebonne Parish, commonly known as Whiskey Island. It is one of three barrier islands that once comprised Isle Derniers, a massive 25-mile-long island that has been eaten away and fragmented over the last century by hurricanes and subsidence. For more than a year, sand was pumped 10 miles to the island by a dredge from Ship Shoal, the massive sand deposit on the Gulf floor that also contributed to the restoration of Elmer's Island. Gradually built up by the Mississippi River about 7,000 years ago, Ship Shoal will be tapped again in 2020 and 2021 for projects at West Belle Pass, Timbalier Island, and Trinity Island, which is immediately east of Whiskey Island. All four projects are addressing the weaknesses in the barrier island shorelines of the Terrebonne Basin, fortifying protections for coastal communities and sustaining critical fish and wildlife habitat—all using funding from oil spill penalties. The $117-million project at Whiskey Island has restored 1,000 acres of beaches and dunes and established another 160 acres of marsh platform. This builds upon the success of a 2009 project that created 300 acres of marsh on the island's east end. Anglers are particularly fortunate that Whiskey Island's beaches and marshes, coated and stained by oil in 2010, have been renewed. The effort has helped to sustain and even enhance Terrebonne's rich recreational and commercial fisheries and give coastal birds—like the brown pelican populations hit hard by the spill—a place to nest and feed for at least two more decades. The project also demonstrates the broader scale of Louisiana's coastal restoration efforts now that oil spill dollars have become available. Past barrier-island restoration efforts were pieced together over a decade or more. But with oil spill penalties committed by Louisiana and federal resource agencies like NOAA and the U.S. Fish and Wildlife Service, larger and more comprehensive projects can be built all at once. This ultimately saves money and makes for a more resilient and productive island. Without the consistent, and now increased, investment in the restoration of Terrebonne Parish's barrier islands over the last 30 years, it's likely those islands would be little more than sandbars and submerged humps today, instead of continuing to be fish and bird meccas and the first-line of defense against hurricanes. Mississippi River Reintroduction at Maurepas Swamp Looking ahead, the restoration of Maurepas Swamp is a unique opportunity for the CPRA to expand coastal habitat restoration and sustainability projects beyond rebuilding coastal marshes and barrier islands. Maurepas Swamp is a massive area of old-growth cypress-tupelo swamp nestled between Baton Rouge and New Orleans that separates Lake Maurepas from the Mississippi River. The swamp was once connected to areas that were flooded annually by the Mississippi, but it was isolated by levees more than a century ago. The natural hydrology of the swamp has also been interrupted by roadbeds and spoil banks from oil and gas canals. Without the annual floods from the Mississippi River allowing sheet water to flow over the swamp, the native trees have lost a vital source of nutrients and the
1.330151
m-a-p/FineFineWeb
Why Toys for Children When you donate a much-needed and treasured toy, you are providing so much more than the toy itself. Here's why: Play is the "work" of young children. From earliest infancy on, toys provide the motivation and the medium for children to learn to control their bodies. Toys help children understand that their bodies are separate from objects, help them learn to reach and grasp, and provide a motivation to learn to roll, sit, crawl, stand, and walk. Through play with toys, children learn concepts of shape, weight, and size; by dropping and throwing toys, they also learn about gravity. The National Institute for Early Education Research suggests that without sufficient toys, children don't develop necessary motor skills. Children learn to use language during play with toys. Studies also show that infant toys are critical for brain growth that affects learning later in life. Toys help prepare children for the academic learning that they will encounter when they enter formal school settings. Toys are also critical to children's emotional development. According to Pamela Schachter, M.S. Ed., University of Medicine and Dentistry of New Jersey Early Intervention Program Manager, "Through toys, children learn that their actions have an impact on the world around them. Playing with a toy with another person encourages shared attention, mutual pleasure and is a child's earliest 'conversation' with loving adults." As children get older they learn about sharing, taking turns, and working collaboratively with others. Toys often serve as emotional security objects for young children, acting as a bridge for the security of a loving caregiver during times of separation. Toys provide opportunities for laughter, fun, and inquisitive exploration of the world. These are critical opportunities for the healthy development of children—emotionally, intellectually, and physically.
2.570434
Zyphra/Zyda-2
Skip over navigation Applying Newton's Three Laws Tension Force How to Cite This SparkNote Problem : A 5kg picture frame is held up by two ropes, each inclined 45 o below vertical, as shown below. What is the tension in each of the ropes? Because the picture frame is at rest, the tension in the two ropes must exactly counteract the gravitational force on the picture frame. Drawing a free body diagram we can calculate the vertical components of the tension in the ropes: Clearly the horizontal components of the tension in the two ropes cancel exactly. In addition, the vertical components are equal in magnitude. Since F = 0, then the vertical components of the tension in the two ropes must cancel exactly with the gravitational force: 2T y = mgâá'2T sin 45 o = (5)(9.8) = 49N. Thus: T = = 34.6N. The total tension on each rope is thus 34.6N. Problem : Consider a 10kg block resting on a frictionless plane inclined 30 o connected by a rope through a pulley to a 10kg block hanging free, as seen in the figure below. What is the direction and magnitude of the resulting acceleration of the 2-block system? Though this problem seems quite complex, it can be solved by simply drawing a free body diagram for each block. Since the resulting acceleration of each block must be of the same magnitude, we will get a set of two equations with two unknowns, T and a. First we draw the free body diagram: On block 1, there are 3 forces acting: normal force, gravitational force and tension. The gravitational force, in terms of parallel and perpendicular components, and the normal force can be easily calculated: F G = (10kg)(9.8) = 98N F Gâä¥ = F Gcos 30 o = 84.9N F G || = F Gsin 30 o = 49N The normal force is simply a reaction to the perpendicular component of the gravitational force. Thus F N = F Gâä¥ = 84.9N. F N and F Gâä¥ thus cancel, and the block is left with a force of 49N down the ramp, and the tension, T, up the ramp. On block 2, there only two forces, the gravitational force and the tension. We know that F G = 98N, and we denote the tension by T. Using Newton's Second Law to combine the forces on block 1 and block 2, we have 2 equations and 2 unknowns, a and T: F = ma 10a 1 = T - 49 10a 2 = 98 - T However, we know that a 1 and a 2 are the same, because the two blocks are bound together by the rope. Thus we can simply equate the right side of the two equations: T - 49 = 98 - T Thus 2T = 147 and T = 73.5N With a defined value for T, we can now plug into one of the two equations to solve for the acceleration of the system: 10a = 73.5 - 49 = 24.5 Thus a = 2.45m/s 2. Interpreting our answer physically, we see that block 1 accelerates up the incline, while block 2 falls, both with the same acceleration of 2.45m/s 2. Problem : Two 10kg blocks are connected by a rope and pulley system, as in the last problem. However, there is now friction between the block and the incline, given by μ s =.5 and μ k =.25. Describe the resulting acceleration. We know from the last problem that block 1 experiences a net force up the incline of 24.5 N. Since friction is present, however, there will be a static frictional force counteracting this motion. F s max = μ s F n = (.5)(84.9) = 42.5N. Because this maximum value for the frictional force exceeds the net force of 24.5 N, the frictional force will counteract the motion of the blocks, and the 2 block system will not move. Thus a = 0 and neither block will move. Follow Us
2.428545
Zyphra/Zyda-2
What would you use a pressure-sensitive keyboard for? This prototype keyboard from Microsoft might change keyboards as we know them. Or it might go over like a Segway running Vista. It all depends on the answer to this question: Is a pressure-sensitive keyboard useful? The video above demonstrates the potential of the technology, showing practical uses like pushing harder on a key in a game to make a character run faster. It can also use the amount of force to determine whether you want to write in lower case or all-caps. My favorite: How hard you hit the Delete key determines whether you want to delete a character or a word. Presumably slamming it as hard as you can will wipe your entire drive. Whether or not that sounds sexy to you, the idea is much closer to reality than most concepts for a few reasons: 1) A working prototype exists. 2) This is Microsoft. 3) The keyboard uses existing tech, with rubber domes underneath every key pushing three flexible membranes. Because of the design, the harder you push, the greater the area of electrical contact, which can be easily measured. Want one of these? What would you use it for? Let us — and Microsoft — know in the comments, and look below for pictures of the prototype. UIST 2009 Student Innovation Contest, via YouTube
1.287379
Zyphra/Zyda-2
next up previous [pdf] Next: 1-D time-varying median filter Up: Theoretical basis Previous: 1-D stationary median filter Signal-to-noise ratio (SNR) estimation using the stack method A simple definition of the SNR was introduced by Liu and Li (1997). Window $D$, a part of the seismic record, can be chosen for SNR analysis: \begin{displaymath} D = [x_{i,j}]_{M \times N} \qquad (0< M \le N_x, 0< N \le N_t)\;. \end{displaymath} (2) Further assumptions are: waveform, amplitude, and phase of seismic wavelet in window $D$ keep stable in respect to distance ''$i$''; noise is ''zero mean'' randomly distributed, along with survey-line direction being independent (decorrelated) of the signal, so that \begin{displaymath} x_{i,j} = s_j + n_{i,j}\; \end{displaymath} (3) \begin{displaymath} \sum_{i=1}^{M} n_{i,j} = 0\;, \end{displaymath} (4) where $s_j$ is amplitude of signal, and $n_{i,j}$ is amplitude of noise. These assumptions generally imply a limitation to this method, but they can be satisfied if the local window is chosen in the stable signal region of the seismic section. So if the signal energy in the window is \begin{displaymath} E_S = M\sum_{j=1}^{N} s_{j}^2 = \frac{1}{M} \sum_{j=1}^{N}(\sum_{i=1}^{M}x_{i,j})^2\;, \end{displaymath} (5) the noise energy can be calculated by \begin{displaymath} E_N = \sum_{j=1}^{N}\! \sum_{i=1}^{M}x_{i,j}^2 - E_S\;. \end{displaymath} (6) Finally, a decibel expression of the SNR is estimated as \begin{displaymath} SNR = \frac{E_S}{E_N} = 10\,\log_{10}{\left(\frac{\displa... ...\sum_{j=1}^{N}\big(\sum_{i=1}^{M}x_{i,j}\big)^2}\right)}\;. \end{displaymath} (7) next up previous [pdf] Next: 1-D time-varying median filter Up: Theoretical basis Previous: 1-D stationary median filter 2013-07-26
1.500763
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
The Online Teacher Resource Receive free lesson plans, printables, and worksheets by email: - Over 400 Pages - Great Writing Habits. - Character Sketches, Plot Summaries - Excellent for students Age Range: Kindergarten through Grade 2 (Early Elementary or Primary Level) Overview and Purpose : In this activity, students imagine what they would like to have fall from the sky every day and describe what happens when too much of it falls at once. Objective: The student will be able to write a short story and draw a picture of what favorite item they would like to have fall from the sky every day. Drawing paper/writing paper Read Cloudy with a Chance of Meatballs to your students. Talk about what happened to the town. Have your students write a short story about what they would like to have fall from the sky everyday and what would happen if it got out of control. When they are finished writing, have them draw a picture of their story. Have the students share their stories with the class when they are finished. This activity can also be done in small groups. The students can decide on one thing they would like to have fall from the sky and then tell their story.
2.558826
HuggingFaceFW/fineweb-edu
What Happens in the Amygdala... Damage to Brain's Decision-Making Area May Encourage Dicey Gambles Individuals with amygdala damage are more likely to lay a risky bet Imagine you've lost your job. You have some money saved, and a chance to double it with a gamble. But if you lose the bet, you'll forfeit everything. What would you do? Most people would not gamble their savings, according to Benedetto De Martino of California Institute of Technology, author of a study published February 8 in Proceedings of the National Academy of Science. People tend to choose avoiding losses over acquiring gains—a behavior known as loss-aversion. But people with damage to the amygdala—an almond-shaped part of the brain involved in emotion and decision-making—are more likely to take bigger risks with smaller potential gains, De Martino's study found. Two women with bilateral amygdala damage showed a dramatic reduction in loss aversion compared with age-matched control subjects on a series of experimental gambles, despite understanding full well the values and risks involved. De Martino already suspected that the amygdala was crucial for loss-aversion based on earlier studies using functional magnetic resonance imaging (fMRI). But these two rare cases with damage to the very structure in question allowed De Martino to directly test his hypothesis. "In functional MRI, you never know if the response is reflecting something else. With the amygdala injury, you have an on "off" response," De Martino says. Whereas healthy control subjects who stood to win $20 but lose $15 were less likely to gamble than if they stood to win $50 but lose only $10, the two patients with amygdala damage were much less affected by the disparity between potential gains and losses. In some cases, they chose to gamble even when the potential losses outweighed the potential gains. The concept of loss-aversion can be applied to many things. Take, for example, someone considering elective surgery. The more serious the possible complications—not matter how unlikely—the less likely he or she is to proceed. But De Martino is especially interested in how loss-aversion applies to money. "My dream would be an economic theory that could capture the complexities of human behavior, based on real people and real brains rather than assumptions," he explains. De Martino's field of study is aptly named neuroeconomics. "It is a more biological view of social science," he says. "A lot of economic theorists think about humans as machines and forget about emotional processing." In previous work, De Martino showed that amygdala damage makes people behave more like these theoretical machines. "Strangely enough, people with damage to the emotional system are, paradoxically, more rational in making certain types of decisions. Their decisions don't take into account any emotional processing." Interestingly, the older of the two women not only lacked loss-aversion, she was even loss-seeking. This difference between the two women was mirrored by differences between their age-matched controls. "As you get older, you get less loss-averse," De Martino says, explaining that even the older control subjects were less fearful of loss. "Your perspective on life changes because you have fewer years to live." This effect could stem from age-related reductions in the volume of the amygdala—as we age, our brains shrink. De Martino says in addition to age, other factors such as income and education are also at play. The amygdala damage did not appear to affect risk-aversion—a similar behavior with an important difference. People who are risk-averse are less likely to take chances even when they have nothing to lose. Whereas loss-aversion might sound abstract to an economist, De Martino says, it probably reflects a very ancient mechanism in the brain. "Think about an animal. It has to get food, but at the same time it has to protect itself from predators. It would be very wise for an animal to weigh gains and losses from an evolutionary perspective." Although a long time has passed since humans had to worry about predators, financial well-being is necessary for survival in the modern world, albeit more abstractly. De Martino's study suggests that the amygdala—known to be involved in processing fear—may make us afraid to risk losing money. "It may be that the amygdala controls a very general biological mechanism for inhibiting risky behavior when outcomes are potentially negative," De Martino says. Rights & Permissions Share this Article: Email this Article
1.51035
Zyphra/Zyda-2
Honor, Privilege, Duty Americans are blessed to have a "government of the people, by the people, and for the people." But democracy has one key requirement: our participation. And the most basic democratic participation is voting. Yet it can also have the most profound impact. When you vote, you help determine who will lead our nation, make our laws and protect our liberties. More than a right, voting is a privilege that millions of people in other parts of the world can only dream about. Still, many Americans choose not to vote. According to the U.S. Census Bureau, as many as 25 to 35 percent of eligible Americans are not registered to vote — that's 45 to 65 million people! Less than half of the voting-age population actually votes in any given election. What's worse, according to the Republican National Committee (from data provided by The Pew Research Center, Voter News Service, and Gallup and Zogby polls), only one out of four eligible evangelical Christians voted in the 2000 elections. Of 59 million potential evangelical voters, 24 million of them were not registered to vote. Only 20 million of the 35 million registered evangelical voters went out and actually voted. 1 The Influence of Faith Author Jim Nelson Black, in his book When Nations Die, wrote: "One of the greatest reasons for the decline of American society over the past century has been the tendency of Christians who have practical solutions to abandon the forum at the first sign of resistance. . . . In the parable of the talents, Jesus Christ did not warn us to run away, to flee to the hills, or hide our eyes, but to go into the fields and bring forth the harvest." 2 Our faith in God should influence our values in life, and that includes the political arena. We shouldn't be bashful about injecting notions of right and wrong into public debates. These ideas come from moral standards, which are essential to a free society such as ours, to prevent it from sliding into social chaos. People of faith, grounded in moral truth, must be prepared to discern those candidates best able to uphold moral values. As Christ's representatives on earth, we are under a mandate to be "salt and light" in our culture (Matthew 5:13–16). Dr. James Dobson wrote, "We live in a representative form of government where we are its leaders. It means that every citizen has a responsibility to participate in the decisions that are made . . . and that includes people of faith . . . [using] his or her influence for what is moral and just." 3 Voting Your Values Here's more sobering news: Many believers fail to consider their biblical values when voting, often choosing candidates whose positions are at odds with their own beliefs, convictions, and values. A recent study by the Pew Forum on Religion and Public Life shows that nearly two-thirds of Americans say their faith has little to do with their voting decisions. But Charles Colson, founder of the Prison Fellowship ministry, highlighted the need for something more than voting the party line when he wrote: Societies are tragically vulnerable when the men and women who compose them lack character. A nation or a culture cannot endure for long unless it is undergirded by common values such as valor, public-spiritedness, respect for others and for the law; it cannot stand unless it is populated by people who will act on motives superior to their own immediate interest. Keeping the law, respecting human life and property, loving one's family, fighting to defend national goals, helping the unfortunate, paying taxes — all these depend on the individual virtues of courage, loyalty, charity, compassion, civility and duty. 4 A Final Word! It is absolutely critical that Christians be registered, be informed and vote our values. As believers, we have a duty to be involved in the democratic process. But our charge is not simply to vote our geographic origins, our denominational affiliation, our party or even our pocketbook. Instead, we must vote our values, our beliefs, our convictions. An informed vote takes effort. The choice is not always easy, but it is always significant. Taking Action: Register to Vote Here are three simple steps: - Use the voter registration form accompanying this brochure. (If you don't have a voter registration form, you can obtain and complete one at _URL_.) - Print. Sign. Seal. Address with label. Stamp. - Mail to your Secretary of State. (To locate the address, see _URL_.) Register Your Friends Help organize a voter registration drive in your local church or community. A special Voter Impact Toolkit is available from Focus on the Family. Call 1.800.A. FAMILY for information. Vote Your Values - Educate Yourself on Election Issues: The website above is a great place to start. - Pick Your Party: Consider how each political party's views on moral issues line up with your core values. - Choose a Candidate: Know what the candidates stand for. Weigh the candidates' values against the life principles found in God's Word, and choose the candidates whose views most closely match. Pray for the Election Election Day, November 2, 2004, is fast approaching, and there is much at stake. In many ways, the future of our nation hangs in the balance. We encourage all concerned Christians to devote themselves to focused prayer for the election and for our nation. A great prayer guide can be found at _URL_. - Candi Cushman, "Remember Florida," Citizen, July 2004, p. 24. - Jim Nelson Black, When Nations Die: America on the Brink: Ten Warning Signs of a Culture in Crisis, (Wheaton, IL: Tyndale House Publishers, 1994), p. xix. - Dr. James Dobson, Focus on the Family Newsletter, June 1996. - Charles Colson, Against the Night: Living in the New Dark Ages, (Ann Arbor, MI: Servant Publications, 1989), p. 67.
1.218977
m-a-p/FineFineWeb
2 edition of Method ad means found in the catalog. Method ad means Written in English Unitary method has been explained here in detail. Click to know what is unitary method, unitary method worksheet, examples, formula and applications along with solved examples. Definition. Means and Methods of Construction — a term used in construction to describe the techniques and tactics (usually temporary structures) a contractor employs to complete construction of a permanent project or structure. This Web SIte. This web site was inspired by Peter Wludyka and Karen Copeland, authors of the book Analysis of Means: A Grpahical Method for Comparing Means, Rates, and goal of this site is to provide an introduction to ANOM methods and more importantly, provide links to research papers, software solutions, and other ANOM resources. Means and methods and repetitive operations drive the contractor's schedule. Means and methods can be influenced by issues such as adverse ground conditions, material deliveries, coordination with adjacent contractors, agency approvals, weather conditions, and many other factors. These issues add uncertainty and risk to the project. Means and Methods of Construction means the labor, materials, temporary structures, tools, plant, and construction equipment, and the manner and time of their use, necessary to accomplish the result intended by this Contract. "Notice to Proceed" or "Order to Work" shall mean the written notice issued by the Commissioner specifying the time for commencement of the Work and the. Explore the research methods terrain, At the most basic level, a codebook describes the layout of the data in the data file and describes what the data codes mean. Codebooks are used to document the values associated with the answer options for a given survey question. Noah Bible Pop-UP Question time in Ceylon Extension of water facilities act. Whose House Is It Anyway Description of the household composition and income files created from the INCAP-Rand Guatemalan survey data Ruffian, international champion. Spoken Kashmiri, a language course Sensation Smith of Drury Lane Patrington monumental inscriptions. The Columbia Guide to American Indians of the Great Plains (The Columbia Guides to American Indian History and Culture) Notes on Sheridans The school for scandal. Methods are often associated with names of individuals who have developed the particular techniques and strategies that they contain. Home Authors, Poets, and Playwrights. The authors survey the history of contextualization in Bible and church history, explore significant proposals of prominent thinkers, and evaluate these proposals from five different perspectives. They conclude their study by outlining their own definition and approach to contextualization, using numerous examples to illustrate their by: The process of price discovery involves generating and recording investor demand for shares before arriving at an issue price. Book building is. Means and Methods Definition. Before delving into Method ad means book main body of the paper and the risks of taking on the traditional contractor roles, following is a short primer on means and methods. The means simply stated are the instruments or equipment used to accomplish something. linear algebra, and the central ideas of direct methods for the numerical solution of dense linear systems as described in standard texts such as,,or. Our approach is to focus on a small number of methods and treat them in depth. Though this book is written in a finite-dimensional setting, we. The Modern Corporation and Private Property is a book written by Adolf Berle and Gardiner Means published in regarding the foundations of Method ad means book States corporate explores the evolution of big business through a legal and economic lens, and argues that in the modern world those who legally have ownership over companies have been separated from their control. RSMeans data is North America's leading construction estimating database available in a variety of formats. Access accurate and up-to-date building construction costs data that helps pre construction managers, architects, engineers, contractors and others to precisely project and control cost estimation of both new building construction and renovation projects. Related concepts. Methodology has several related concepts: paradigm, algorithm, and method. The methodology is the general research strategy that outlines the way in which research is to be undertaken and, among other things, identifies the methods to be used in it. These methods, described in the methodology, define the means or modes of data collection or, sometimes, how a specific result. School for International Training (SIT) in Brattleboro, Vermont and appear in her book. Techniques and Principles in Language Teaching (Oxford ). The six studio instructors, each a specialist in the method/approach being demonstrated, are staff members. The method of offering shares by providing a price range is called as book building method. Book building. Book building is actually a price discovery method. In this method, the company doesn't fix up a particular price for the shares, but instead gives a price range, e.g. Rs About the Book: This second edition has been thoroughly revised and updated and efforts have been made to enhance the usefulness of the book. In this edition a new chapter The Computer: Its Role in Research have been added keeping in view of the fact tha4/5(12). With a method, there are prescribed objectives, roles for teacher and learners and guidelines for activities, and, consequently, little flexibility for teachers in how the method is used. The teacher's role is to implement the method. Audiolingualism, Total Physical Response and Silent Way are examples of methods. Mechanical Estimating Methods. Item# B. An easy-to-use tool for estimating heating, ventilating, and air conditioning systems, with up-to-date cost data and estimating examples. This all-in-one reference gives you the accepted standards and procedures for takeoff and pricing HVAC systems, as well as piping, plumbing, and fire protection. Nothing less than a major contribution to the scientific culture of this world." — The New York Times Book Review This major survey of mathematics, featuring the work of 18 outstanding Russian mathematicians and including material on both elementary and advanced levels, encompasses 20 prime subject areas in mathematics in terms of their simple origins and their subsequent sophisticated Reviews: Qualitative research is a type of social science research that collects and works with non-numerical data and that seeks to interpret meaning from these data that help understand social life through the study of targeted populations or places. uses procedures, methods and techniques that have been tested for their validity and reliability; 3. is designed to be unbiased and objective. Philosophies means approaches e.g. qualitative, quantitative and the academic discipline in which you have been trained. Validity means that correct procedures have been applied to find answers to a. Qualitative research methods is defined as a process that focuses on obtaining data through open-ended and conversational communication. This method is not only about "what" people think but also "why" they think so. Also, read in this blog qualitative research examples, types, & more. Look and Say Teaching Method The look and say teaching method, also known as the whole word method, was invented in the s and soon became a popular method for teaching reading. By the s and s there was a very strong focus on teaching children to read by this method. Search the world's most comprehensive index of full-text books. My library. What does method mean. method is defined by the lexicographers at Oxford Dictionaries as A particular procedure for accomplishing or approaching something, especially a systematic or. Classification; a mode or system of classifying natural objects according to certain common characteristics; as, the method of Theophrastus; the method of Ray; the Linnæan method. Method (noun) A technique used in acting in which the actor tries to identify with the individual personality of the specific character being portrayed, so as to provide a realistic rendering of the character's role.method: [noun] a procedure or process for attaining an object: such as. a systematic procedure, technique, or mode of inquiry employed by or proper to a particular discipline or art. a systematic plan followed in presenting material for instruction. a way, technique, or process of or for doing something. a body of skills or techniques. Accounting method refers to the rules a company follows in reporting revenues and expenses. The two primary methods are accrual accounting and cash accounting.
2.073087
openbmb/Ultra-FineWeb
One hormone out of balance can disrupt and prevent pregnancy. Canary Club's at home Fertility Hormone Tests can help guide you and your health practitioner in discovering hormone-related fertility issues. Let's look at two of the phases of the women's cycle: the Follicular Phase and Luteal Phase, where pregnancies have their most common complications including miscarriages. At its most simplified, pregnancy involves these five basic steps. Hormones interact at every level, progressing each step and moving the cycles forward.Fertility Cycle chart 1. Follicular Phase - making an egg 2. Ovulation - releasing the egg 3. Fertilization 4. Implantation 5. Gestation Making An Egg The follicular phase is the selection of one egg from thousands within the ovaries to be the "dominant" egg for that month. After 14 days, this dominant egg will be released. After menstruation, which occurs during the first few days of the follicular phase, the estradiol levels start to rise causing the growth of the endometrial lining and maturation of the egg and the follicle. Release the Egg (Ovulation) At ovulation, a surge of luteinizing and follicle-stimulating hormones stimulate the release of the egg from the follicle and the tissue left behind becomes the corpus luteum. If the egg is not fertilized it will die and progesterone and estrogen levels fall. Without these hormones, the uterus lining breaks up and is released in menstruation. However, if the egg IS fertilized, it becomes a "blastocyst" and will make its attempt at implanting in the uterus. Implantation typically occurs between 6-10 days after ovulation. Upon successful implantation, the now embryo produces the hormone Human Chorionic Gonadotrophin (HCG) which tells the ovaries to continue to produce progesterone. Progesterone suppresses the surge of estrogen that typically happens just before the release of an ovum, which in turn inhibits the production of Luteinizing Hormone (LH) and Follicle-Stimulating Hormone (FSH). This chain reaction of hormone production stops the release of more eggs during pregnancy. Home Fertility Test, Fertility Hormone Testing Eventually, during the 10th to 12th week of gestation, the placenta relieves the corpus luteum's job and begins to produce progesterone, which it continues to do throughout the pregnancy. Complications to Pregnancy and Miscarriage An imbalance of hormones during the gestation phase can cause miscarriages. This can be a result of a failure of the follicle to develop due to inadequate follicle-stimulating hormone, or due to the early breakdown of the corpus luteum. Either way, a defective corpus luteum results in inadequate progesterone production and is one of the leading causes of early termination of pregnancy (miscarriage) and may also be involved in the failure of an embryo to implant in the first place. Achieving and Maintaining a Pregnancy ZRT Fertility TestOrder an easy to use, at home fertility test today! Comprehensive Fertility Test Menstrual Cycle Mapping Test Basic Reproductive Hormone Trio Test Single hormone tests (good for affordable re-testing): Luteinizing Hormone (LH) Follicle-Stimulating Hormone (FSH) The International Council on Infertility Information Dissemination, Inc. Full Script Supplements healthbalance coach fullscript 4 200x200
2.25433
Zyphra/Zyda-2
Rush ladies, rush; before someone else or the lady next door gets there.The astronomers have discovered a new planet made of all diamond. The new planet is far denser than any other known so far and consists largely of carbon. Because it is so dense, scientists calculate the carbon must be crystalline, so a large part of this strange world will effectively be diamond. Lying 4,000 light years away, or around an eighth of the way toward the center of the Milky Way from the Earth, the planet is probably the remnant of a once-massive star that has lost its outer layers to the so-called pulsar star it orbits. What impact will this new discovery have on the mankind who must have started feeling uncomfortable after hearing the news? If they had the necessary means, "the diamond rush" would already have started taking place but alas! the new planet is far away and there are no known means to reach there. This reminds of the famous California Gold Rush (1848–1855) began on January 24, 1848, when gold was found by James W. Marshall at Sutter's Mill in Coloma, California. The first to hear confirmed information of the gold rush were the people in Oregon, the Sandwich Islands (Hawaii), and Latin America, and they were the first to start flocking to the state in late 1848. All in all, the news of gold brought some 300,000 people to California from the rest of the United States and abroad. Of the 300,000, approximately half arrived by sea and half came overland on the California Trail and the Gila River trail. The gold-seekers, called "forty-niners" (as a reference to 1849), often faced substantial hardships on the trip. While most of the newly arrived were Americans, the Gold Rush attracted tens of thousands from Latin America, Europe, Australia, and even China. At first, the prospectors retrieved the gold from streams and riverbeds using simple techniques, such as panning. More sophisticated methods of gold recovery were developed and later adopted around the world. At its peak, technological advances reached a point where significant financing was required, increasing the proportion of gold companies to individual miners. Gold worth tens of billions of today's dollars was recovered, which led to great wealth for a few. However, many returned home with little more than they had started with. The effects of the Gold Rush were substantial. San Francisco grew from a small settlement of about 200 residents in 1846 to a boomtown of about 36,000 by 1852. Roads, churches, schools and other towns were built throughout California. In 1849 a state constitution was written, a governor and legislature chosen and California became a state in 1850 as part of the Compromise of 1850. New methods of transportation developed as steamships came into regular service. By 1869 railroads were built across the country from California to the eastern United States. Agriculture and ranching expanded throughout the state to meet the needs of the settlers. At the beginning of the Gold Rush, there was no law regarding property rights in the goldfields and a system of "staking claims" was developed. The Gold Rush also had negative effects: Native Americans were attacked and pushed off their lands and the mining has caused environmental harm. An estimated 100,000 California Indians died between 1848 and 1868 as a result of American immigration. The physical coordinates of the planet do not suggest the possibility of steamship or railroad a la California Gold Rush. There is also no possibility of population surge and no possible danger to the lives of natives, if there are any. The measurements of the newly-discovered planet suggest the planet, which orbits its star every two hours and 10 minutes, has slightly more mass than Jupiter but is 20 times as dense. In addition to carbon, the new planet is also likely to contain oxygen, which may be more prevalent at the surface and is probably increasingly rare toward the carbon-rich center. Its high density suggests the lighter elements of hydrogen and helium, which are the main constituents of gas giants like Jupiter, are not present. Just what this weird diamond world is actually like close up, however, is a mystery.
2.079453
openbmb/Ultra-FineWeb
In ancient Greek democracies (e.g., Athens) public officials were occasionally elected but more often were chosen by lot. In Rome the popular assemblies elected the tribunes. In the Middle Ages elections were abandoned, except for such processes as elections to the papacy and, in a more limited sense, of the Holy Roman emperor by a small and partly hereditary body of electors. In the modern period, elections have been inseparable from the growth of democratic forms of government. Elections were associated with the parliamentary process in England from the 13th cent. and were gradually regularized by acts prescribing the frequency of elections (the Triennial Act of 1694, and the Septennial Act of 1716), by successive reform bills widening the franchise in the 19th cent., and by the adoption of the secret ballot in 1872. In colonial America the election of church and public officials dates almost from the founding of the Plymouth Colony, and the paper ballot was instituted in elections to the Massachusetts governorship in 1634. Under the U.S. Constitution the right to hold elections is specified, but the method and place are left to the states, with Congress having the power to alter their regulations. The Constitution specified that elections to the House of Representatives be direct, or popular, and that the election of the Senate and of the president and vice president be indirect, Senators being chosen by the state legislatures and the president and vice president by electors selected by the people (see electoral college). The Seventeenth Amendment (1913) provided for popular election of senators. Political candidates are usually chosen by delegate convention, direct primary, nonpartisan primaries, or petition. The candidate who receives the most votes is usually elected, but an absolute majority may be required; a majority has not been required in the U.S. federal elections since 1850 except in the electoral vote cast for the president and vice president. (In presidential nominating conventions an absolute majority is required; the Democrats required a two-thirds vote of the delegates from 1832 to 1936.) Since 1824, when John Quincy Adams was elected president by the House of Representatives when no candidate had a majority in the electoral college (Adams' leading opponent, Andrew Jackson, had greater popular and electoral vote totals), the electoral-college system has many times permitted a president to be chosen without a majority of the popular vote (1844, 1848, 1856, 1860, 1876, 1880, 1884, 1888, 1892, 1912, 1916, 1948, 1960, 1968, 1992, 1996, and 2000); in 1876, 1888, and 2000 candidates without even a plurality succeeded in winning office. Voting frauds and disorder at the polls were common after the rise of political machines, and the enactment of registration laws after 1865 did little to ameliorate conditions. Corrupt practices acts, poll watching, the institution of primary elections, and the introduction of voting machines after 1892 have been more effective in ensuring honest elections. Nonetheless, the disputes over the counting of the votes in Florida in the 2000 presidential election clearly revealed that some machine voting systems are more reliable than others and that less reliable systems can potentially distort the results. All states have some residency requirements as a condition for suffrage. Originally the vote was typically restricted to white men who met certain property qualifications; such restrictions were later liberalized and removed. The Fourteenth Amendment (1868) and Fifteenth Amendment (1870) were designed to forbid the disenfranchisement of African-American men after the Civil War, and the Nineteenth (1920) conferred the vote on women. The Twenty-third Amendment (1961) permitted residents of the District of Columbia to vote in the presidential elections, while the Twenty-fourth Amendment (1964) outlawed payment of poll or other taxes as a condition for voting. The Twenty-sixth Amendment (1971) lowered the voting age from 21 to 18. The so-called motor voter law (National Voter Registration Act, 1993) was designed to reverse declining voter registrations by permitting registration at motor vehicle departments and other agencies. Certain classes of felons and some others are deprived of the vote. In the United States, Canada, Mexico, Great Britain, and many other nations, usually the one candidate who receives the most votes is elected from a district. If an absolute majority is needed to win, a run-off election maybe required. In Denmark, Norway, Sweden, Russia, Ukraine, Brazil, and other nations (including some local elections in the United States), proportional representation is used to more adequately represent political or other minorities within the electorate. Two methods are used—the voter ranking system in a multiseat district, and the party list system in which parties gaining a threshold vote win one seat and those well above the threshold dividing the remaining seats proportionally. While a proportional system provides representation for minorities and eliminates gerrymandering, primaries, and run-off elections, it can weaken the representative-constituent relationship, encourage multiple parties, and necessitate coalition rather than majority governments. Some nations, for example, Italy, Germany, and Spain, use a combination of direct election and proportional representation. See E. Lakeman, How Democracies Vote (1970); E. H. Rosebloom, A History of Presidential Elections (3d ed. 1970); H. A. Bone, American Politics and the Party System (4th ed. 1971) and Politics and Voters (3d ed. 1971); J. M. Clubb, ed., Electoral Change and Stability in American Political History (1971); J. H. Silbey et al., The History of American Electoral Behavior (1978); S. A. and B. G. Samore, Candidates, Parties, and Campaigns (1985); W. R. Neuman, The Paradox of Mass Politics (1986). Electoral device for choosing a party's candidates for public office. The formal primary system is peculiar to the U.S., where it came into widespread use in the early 20th century. Most U.S. states use it for elections to statewide offices and to the national presidency; in presidential elections, delegates are selected to attend a national convention, where they vote for the candidate to whom they are pledged. A closed-vote primary is restricted to party members; an open-vote primary is open to all voters in the district. Names can be placed on a ballot by an eligible citizen's declaration of candidacy, by nomination at a pre-primary convention, or by a petition signed by a required number of voters. In the late 20th and early 21st centuries political parties in some countries (e.g., the United Kingdom and Israel) adopted similar procedures for the election of the national party leader. Seealso electoral system; party system. Learn more about primary election with a free trial on Britannica.com. Formal process by which voters make their political choices on public issues or candidates for public office. The use of elections in the modern era dates to the emergence of representative government in Europe and North America since the 17th century. Regular elections serve to hold leaders accountable for their performance and permit an exchange of influence between the governors and the governed. The availability of alternatives is a necessary condition. Votes may be secret or public. Seealso electoral system, party system, plebiscite, primary election, referendum and initiative. Learn more about election with a free trial on Britannica.com. An election is a decision-making process by which a population chooses an individual to hold formal office. This is the usual mechanism by which modern democracy fills offices in the legislature, sometimes in the executive and judiciary, and for regional and local government. This process is also used in many other private and business organizations, from clubs to voluntary associations and corporations. The universal acceptance of elections as a tool for selecting representatives in modern democracies is in contrast with the practice in the democratic archetype, ancient Athens. Elections were considered an oligarchic institution and most political offices were filled using sortition, also known as allotment, by which officeholders were chosen by lot. Electoral reform describes the process of introducing fair electoral systems where they are not in place, or improving the fairness or effectiveness of existing systems. Psephology is the study of results and other statistics relating to elections (especially with a view to predicting future results). Historically, other groups of people have also been excluded from voting. For instance, the democracy of ancient Athens did not allow women, foreigners, or slaves to vote, and the original United States Constitution left the topic of suffrage to the states; usually only white male property owners were able to vote. Much of the history of elections involves the effort to promote suffrage for excluded groups. The women's suffrage movement gave women in many countries the right to vote, and securing the right to vote freely was a major goal of the American civil rights movement. Extending the right to vote to other groups which remain excluded in some places (such as convicted felons, members of certain minorities, and the economically disadvantaged) continues to be a significant goal of voting rights advocates. Suffrage is typically only for citizens of the country. Further limits may be imposed: for example, in Kuwait, only people who have been citizens since 1920 or their descendants are allowed to vote, a condition that the majority of residents do not fulfill. However, in the European Union, one can vote in municipal elections if one lives in the municipality and is an EU citizen; the nationality of the country of residence is not required. In some countries, voting is required by law;
2.007952
HuggingFaceFW/fineweb-edu
- About Extension - Get Involved - Statewide Locations Why do flowers drop off veggie plants? This article has been updated. Please check our website for the most recent story. May 14, 2004 CORVALLIS - Do the flowers of your green beans, cucumbers, melons, squash, tomatoes, eggplant or peppers bloom, and then drop off without setting fruit? Or do they not set bloom at all? Some of the reasons behind blossom drop or too few blossoms on your vegetable plants may include too much heat, lack of pollination, too cold and water stress. Deborah Kean, researcher at Oregon State University's Vegetable Research Farm in Corvallis, gives home gardeners some background on blossom drop in major groups of vegetables. SQUASH – Gardeners tend to notice blossom drop in squash because squash flowers are so huge and noticeable. Squash have separate male and female flowers. The males always drop off after they shed their pollen. Female flowers may fall off due to lack of pollination. Blossom drop is often associated with cool temperatures, as bees and other pollinators aren't out collecting and transferring pollen as much when temperatures are down. Sometimes female flowers will bloom before there are any male flowers out. "Typically, a few female flowers will bloom before there are any male flowers, so, of course, they don't get pollinated," explained Kean. "Because female flowers already have what looks like a baby fruit on the end (inferior ovary, as botanists call it), people think the fruit is aborting when, in fact, it is just unpollinated flowers drying up and falling off." If you want to help Mother Nature along, you can hand pollinate squash by removing a male flower, then rub the pollen producing anthers (top of the stamen) onto the stigma of the female flower. TOMATOES AND PEPPERS, EGGPLANTS – In areas of Oregon where late spring and summer nights are cooler, blossom drop in tomatoes and peppers can occur. Too much nitrogen and water can stimulate a lot of leafy growth, but few blossoms. Heat can cause blossom drop as well. Tomato blossoms drop off without setting fruit when night temperatures fall below 55 degrees or day temperatures exceed 90 degrees for extended periods. Cherry tomatoes and the OSU-bred parthenocarpic tomatoes, including Gold Nugget, Oregon Spring, Oregon Star, Siletz and Legend, are the exception, as they will set fruit over a wider temperature range than most large-fruited types. Parthenocarpic and cherry tomatoes will fruit throughout the heat of summer, even in Tucson, according to information from the University of Arizona. Blossoms drop off if your tomato, pepper or eggplant plants are water stressed. Deep and infrequent watering is best. Also, make sure your plants are getting enough sun. If they receive less than 6 hours of sunlight a day, they may not bloom. Also, windy conditions can interfere with fruit set, so if you can create windbreaks in the garden, do so. Pepper blossoms are even more sensitive to temperature fluctuations during pollination. Normal pollination and fruit set and don't occur when night temperatures fall below 58 degrees and daytime temperatures rise above 85 degrees. Under these temperature conditions, the blossoms fall off, often before pollination. BEANS AND PEAS – Intense heat (generally higher than 95 degrees) or water stress (e.g. under watering) can cause beans and peas to drop their flowers. Snap beans produce way more blossoms than they can set pods, because bean breeders have selected for concentrated pod set. Sometimes, during a heat wave, beans will drop blossoms, sometimes resulting in what growers call a "split set," or two distinct crops of beans in one set, explained Kean. "This drives vegetable processors crazy, as they do a one time harvest," she said. "But this may not bother home gardeners." CUCUMBERS, MELONS – Cooler weather keeps pollinators from flying. If female flowers are not pollinated, they will abort. Male flowers always drop, as they don't form the fruit; they provide the pollen. CORN - Corn is another vegetable adversely affected by high temperatures. The corn tassel (pollen producing part of the corn flower) is often killed when temperatures are above 100 degrees. Injury to the tassel will prevent development of kernels in the ear. Fortunately for most parts of the state, summer temperatures don't stay above 100 degrees for long. Vegetable varieties vary in their sensitivity to blossom drop. If you have trouble with one variety, try a different one next year. By planting varieties adapted to the growing conditions in your region, you can avoid many of the blossom drop and non-pollination problems. And remember, some flower drop on these garden plants is normal. Plants generally produce way more flowers than the number of fruits than they can support. Source: Deborah Kean
1.892532
openbmb/Ultra-FineWeb
- How does art improve focus? - What skills does art develop? - How does art influence life? - How is art beneficial to students? - How does art relieve stress? - How drawing can improve your memory? - What are the positive effects of art? - Is creativity a sign of intelligence? - Can Drawing make you smarter? - Why is art important in life? - Why is art so important? - How does art heal? - Why is drawing so fun? - Does art increase intelligence? - What is the purpose of making art? - Is drawing a talent or a skill? - Why do I like to draw? - How does art affect mental health? - How could I increase my IQ? - Why is drawing so hard? - How does art benefit the human brain? - How can art be smart? - Does art improve memory? - Does drawing make you happy? - How does art make you feel? How does art improve focus? Creating art increases the feel-good neurotransmitter dopamine. Dopamine has been called the "motivation molecule." It boosts drive, focus, and concentration. It enables you to plan ahead and resist impulses so that you can achieve your goals.. What skills does art develop? Skills developed through participation in the arts are increasingly important in the workplace and therefore, key to a successful career.CREATIVITY. … CONFIDENCE. … PROBLEM SOLVING.PERSEVERANCE. … FOCUS. … NON-VERBAL.RECEIVING.COLLABORATION DEVELOPING.More items… How does art influence life? With the art that we are surrounded by, whether it's a painting, music or even videos can have a huge impact on our mood and emotions. … All kinds of art can affect our mood in a positive way, making us feel happier, calmer, or even inspired to do something. Everywhere you go art is evident. How is art beneficial to students? Art programs have the potential to help improve a student's memory and concentration skills, develop decision-making and critical thinking skills, enhance communication and listening skills, but also encourage focus and discipline. … An arts education can also help a student to improve self-esteem and social skills. How does art relieve stress? With art, you have the additional benefit of being left with something beautiful (or at least interesting) to show for it. Helps you tap into a "state of flow": Some psychologists describe flow as becoming deeply engrossed in an activity. Similar to meditation, flow can improve performance and lower stress levels. How drawing can improve your memory? Summary: Researchers from the University of Waterloo found that even if people weren't good at it, drawing, as a method to help retain new information, was better than re-writing notes, visualization exercises or passively looking at images. What are the positive effects of art? The Positive Power of Arthas a positive impact on our sense of hope, self-worth, and well-being.improves our sense of connectedness and widens our social networks.decreases depression and anxiety and reduces stress. Is creativity a sign of intelligence? 1. You're creative. Dr. Katie Davis, a clinical neuropsychologist, told Business Insider that creativity is a definitive sign of intelligence because it requires thinking flexibly/outside the box and necessitates the ability to shift and change your patterns of thinking from one way to another. Can Drawing make you smarter? He believes that drawing and painting helps growing new brain cells. This is not only applicable to patients but in a normal person, drawing actually adds synapses to the brain's transmitters. This means that the memories and experiences reserved in your brain are stronger, more striking, and more accessible. Why is art important in life? Art surrounds life, all people in every location, without us being aware of it. Since time immemorial, art has existed as long as man has. It is a huge part of our culture, which shapes our ideas, and vice versa, and provides us with a deeper understanding of emotions, self-awareness, and more. Why is art so important? Art is important because it encompasses all the developmental domains in child development. Art lends itself to physical development and the enhancement of fine and gross motor skills. … Children learn about themselves and others through art activities. It really helps them build self-esteem. How does art heal? A: Art is healing because it forces you to forge a connection between your mind and your body. Unlike exercise, which works your body, or meditation, which clears your mind, art-making accesses both mind and body to promote healing. Why is drawing so fun? Drawing is great fun when you do it as a means of exploration; of your own mind, the medium, the concepts of form etc, however it can be a dread as well, in particular when you try to draw complex images without having the appropriate technique. Example: doodling is great fun. Does art increase intelligence? Art lessons increase brain plasticity, fluid intelligence, IQ, and attention. They improve overall behavior and reduce impulsiveness. A study of over 10,000 students found that a one-hour trip to an art museum changed the way they thought and felt. What is the purpose of making art? We make art because there is something inside the creative person that needs to get out. The poet, musician, actor, and visual artist all have a desire to express what they feel and to create something of great value. It's a type of therapy or a form of meditation. Many do art for the pure joy of it. Is drawing a talent or a skill? It is both. A skill is something taught and learned, and through levels of learning, you develop expertise. A talent is an inborn and natural inherent ability. The ability to draw will come naturally for some people, and their natural ability can be enhanced by further training and developing higher skill. Why do I like to draw? Drawing helps me discover, reflect on, and express myself. It helps me feel more happy and confident, and has given me an extra sense of identity and purpose in life. It keeps me honest with myself. It's how I grow. It helps me discover my voice. How does art affect mental health? The arts valuable role in mental health is being recognised It can help to boost confidence and make us feel more engaged and resilient. Besides these benefits, art engagement also alleviates anxiety, depression and stress. How could I increase my IQ? Here are some activities you can do to improve various areas of your intelligence, from reasoning and planning to problem-solving and more.Memory activities. … Executive control activities. … Visuospatial reasoning activities. … Relational skills. … Musical instruments. … New languages. … Frequent reading. … Continued education. Why is drawing so hard? The pattern our eyes take when we're drawing something is sequential. … But instead of our eyes flicking all around the object we're drawing, a typical zigzag pattern, they creep around the outline a bit at a time, pausing as they go. That's what makes drawing accurately so hard. How does art benefit the human brain? Making art invigorates the brain in ways that are distinct from merely viewing art. Studies have credited the production of visual art with increases in functional connectivity in the brain along with enhanced activation of the visual cortex. How can art be smart? Art is food for the brain Even the slightest exposure to art can make a person smart, because it can bring in cognitive benefits. It is through exposure to art that a person can begin to think critically. It requires careful decision making, and that increases a person's intelligence. Does art improve memory? People often choose to display art in their home for aesthetic reasons, but recent studies have shown that engaging with the visual arts can actually improve stress, memory, and empathy, whether by viewing art or creating it. Does drawing make you happy? Studies show that art can make you happier – whether you're creating you own art, or enjoying someone else's. … A study by neurobiologist and University College London professor Semir Zeki found that looking at a work of art can actually have the same psychological effect as the euphoric experience of romantic love. How does art make you feel? Art forms give humans a higher satisfaction in emotional release than simply managing emotions on their own. Art allows people to have a cathartic release of pent-up emotions either by creating work or by witnessing and pseudo-experiencing what they see in front of them.
2.079357
openbmb/Ultra-FineWeb
The salutary action of melatonin and betaine, given singly or concomitantly, on cisplatin-induced nephrotoxicity in mice Mohammed Al Za'abi*, Haytham Ali, Mohammed Al Sabahi, Badreldin H. Ali *المؤلف المقابل لهذا العمل نتاج البحث: المساهمة في مجلةمراجعة النظراء 3 اقتباسات (Scopus) ملخص Cisplatin (CP) is commonly used in the treatment of various solid tumors. Its use, however, is hampered by nephrotoxicity. In this study, we compared the effect of betaine and melatonin given singly, with that of a combination of these two agents on CP-induced nephrotoxicity in mice. CP (20 mg/kg, given intraperitoneally on the 8th day of 12 days of the experiment) showed the typical physiological, biochemical, and histologic features of nephrotoxicity. CP-treated mice showed a significant reduction in food intake, body weight, and urine and fecal output. It also induced significant increases in the plasma concentrations of urea, creatinine, uric acid, phosphorous, adiponectin, interleukin-1β, interleukin-6, transforming growth factor -β1, tumor necrosis factor-α, and cystatin C. All these effects were significantly reduced by daily administration of betaine or melatonin at oral doses of 200 mg/kg and 10 mg/kg, respectively. Furthermore, using the two agents in combination caused further significant reductions in the above parameters. These findings suggest that betaine and melatonin concomitant use is likely to provide greater protection against CP-induced nephrotoxicity than when they are given singly, rendering them potentially suitable and safe agents to use in clinical trials to assess their possible beneficial actions in cancer patients receiving CP. اللغة الأصليةEnglish الصفحات (من إلى)1693-1701 عدد الصفحات9 دوريةNaunyn-Schmiedeberg's Archives of Pharmacology مستوى الصوت394 رقم الإصدار8 المعرِّفات الرقمية للأشياء حالة النشرPublished - أغسطس 2021 ASJC Scopus subject areas - ???subjectarea.asjc.3000.3004??? بصمة أدرس بدقة موضوعات البحث "The salutary action of melatonin and betaine, given singly or concomitantly, on cisplatin-induced nephrotoxicity in mice'. فهما يشكلان معًا بصمة فريدة. قم بذكر هذا
1.399287
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
illustrating Science Fiction. Here's an Ancient Novel About Aliens, UFOs, and Robots Written in the 2nd Century The man who introduced science fiction to society. Nearly two millennia before Jules Verne and H. G. Wells started writing about science fiction, Lucian of Samosata, a Syrian satirist, and rhetorician, introduced the world to science fiction, envisioning a future where humankind would find alien life, take place in interplanetary warfare, and come to know artificial life among other things. Lucian of Samosata is widely recognized as one of the earliest novelists in Western Civilization, and many scholars argue he is the father of science fiction. Lucian, a Syrian writer who penned down his works in the Greek language, lived from the year 125 and covered various topics such as extraterrestrials, spaceships, and robots, two millennia before the novels of Jules Verne or G.H. Wells appeared. He developed stories in the key of what today, two millennia later, we conceive as science-fiction. The birth of science fiction Lucian narrates the adventures of Lucian and his crew, all of who were cosmonauts, flying ships, television projections, talking machines, artificial intelligence, encounters with aliens, space battles, ufos, humanoids, as well as cities erected inside a living organism. Illustration from 1894 by Aubrey Vincent Beardsley depicting a battle scene from Book One of Lucian's novel A True Story. Image Credit: Wikimedia Commons / Public Domain. Illustration from 1894 by Aubrey Vincent Beardsley depicting a battle scene from Book One of Lucian of Samosata's novel A True Story. Image Credit: Wikimedia Commons / Public Domain. His work, "A True Story," is a fictional narrative in which he anticipated modern science fiction themes. The work begins with Lucian and his fellow travelers journeying out past the Pillars of Heracles, after which they are blown off course due to a powerful storm. They end up on an island that has a river of wine filled with fish and bears. They don't remain long on the island and make their way when captured by a whirlwind, which ends up taking them on to the moon. There, they find themselves in the middle of a war between the rulers of the moon and the sun, who are fighting a cosmic battle over the colonization of the 'Morning Star.' Both armies are described as having bizarre forms of life, half mechanic, half biological. The armies of the Sun eventually come out on top, after which a peace treaty is signed. In his work, Lucian details how life on Earth is different from that on Earth. Eventually, Lucian and his fellow travelers are swallowed by a 200-mile-long whale as they return to Earth. Inside the creature's belly, they find a variety of fish people inhabiting its inside. They wage war against the mysterious creatures and eventually kill the whale after starting a bonfire in its interior. Ultimately, the manage to open the whale's mouth and escape. As they continue their voyage, they encounter a massive abyss in the ocean but manage to travel beside it. This leads them to discover a distant continent that they decide to explore. The writing ends abruptly with Lucian stating that their future adventures will be described in the upcoming sequels, none of which were ever published. Lucian's work also describes adventures and stories of incredible shipwrecks; He wrote about ships sailing through extraordinary worlds, like an island of dreams, or fantastically wandering through forests and landscapes in space. Lucian is noted for being the first writer to separate reality from fiction, which was not very clear at that time. It is noteworthy to mention that Lucian's book A True Story is also an early expression of the idea of traveling across the Atlantic Ocean and exploring the lands that are located on the other side. This idea predates the journey of Columbus by around 1,400 years. You can read the novel by clicking here. Written by Ivan Petricevic Write for us Get In Touch
2.178408
Zyphra/Zyda-2
The Non-Conforming American – 1955 – Past Daily Reference Room Click on the link here for Audio Player: The Non-Conforming American – March 12, 1955 During the era of the 1950s a gradual, unmistakable undercurrent of social upheaval was starting to take place. Brought on largely by the Cold War, the threat of Nuclear annihilation, the stringent social mores and arch conservatism that swept the country during the Eisenhower years and the beginnings of protest movements, most notably Civil Rights and the seismic social upheaval in the form of Rock n' Roll. All these elements combined to cause some people to question just what was so important about maintaining uniformity, achieving status and material wealth; about not questioning our value system. And so the whole question of non-conformity became the topic of talk shows and magazine articles. In short; what did these people want? Who were they? Where did they come from? Was it a good thing? What exactly was non-conformity? On this radio program, Conversation – which billed itself as "dedicated to the art of good talk", host Clifton Fadiman asked a panel, which included Adlai Stevenson and Alistair Cooke what the definition of non-conformity was and what was it's affect on society? Originally broadcast on March 12, 1955, it seems rather strange now that this subject should be taken so seriously. But fifty-eight years ago it was a different world. The panel struggles in places, not quite sure if non-conformity is the by-product of protest or an emotional/mental problem. What is interesting is how prevalent this shift in our thinking became over the coming years. It's clearly evident even in advertising of the period, which slowly gave way from the traditional and conforming stereotypes to non-conforming messages and brash imagery. And how it grew and continued, and even flourished in what were commonly referred to as "fearful times". And if you think the social/sexual explosion of the 1960s just happened overnight – think again. It was a decade in brewing. - 5 Ways To Practice Non-Conformity In The Matrix (zen-haven.com) - Romanticism's Claim on Individuality (thomascotterill.wordpress.com) - Archives: The Days of Fins and Rockets (motortrend.com) - The Way We Were: 1955 (motortrend.com) ~ by gordonskene on March 13, 2013. Posted in 1950's, Americana, Beat Generation, Broadcasts, Civil Rights, Civil unrest, Cold War, Commentators, Conflicts, Corporate America, Counterculture, Day in history, Disconnect, Disputes, Eisenhower, Generation Gap, Interviews, Mc Carthy era, Mid-Century America, Middle Class, Past Daily, Past Daily Reference Room, Politics, Pop Music, Popular Culture, Post World War 2, Protest Movements, Quality of Life, Racial unrest, Radio, Red Scare, Rock n' Roll, Social Commentary, Society, Uncategorized, Weekend Talkshows Past Tags: Alistair Cooke, Behavior, Clifton Fadiman, Cold War, Conformity, Dwight D. Eisenhower, Nonconformism, United States
1.382143
HuggingFaceFW/fineweb-edu
The spread of English around the world has been and continues to be both rapid and unpredictable. World Englishes: The Study of New Linguistic Varieties deals with this inescapable result of colonisation and globalisation from a social and linguistic perspective. The main focus of the book is on the second-language varieties of English that have developed in the former British colonies of East and West Africa, the Caribbean, South and South-East Asia. The book provides a historical overview of the common circumstances that gave rise to these varieties, and a detailed account of their recurrent similarities in structure, patterns of usage, vocabulary and accents. Also discussed are debates about language in education, the rise of English in China and Western Europe, and other current developments in a world of global travel and migration. 1. History: The spread of English; 2. Structural features of New Englishes I: Morphology and phrasal syntax; 3. Structural features of New Englishes II: Cross-clausal syntax and syntactic theory; 4. More on structure: Lexis, phonology, and optimality syntax; 5. Pragmatics and discourse; 6. Language contact and language acquisition issues in New English research; 7. Conclusion: Current trends in the spread of English.
2.507311
openbmb/Ultra-FineWeb
The Dead Sea and the Land of Israel as It's seen from Nevo Mountain. Rabbi Mira Raz lingers on the moment where Moses asks God if he can see the good land. What is Moses' view? And how does the same fateful moment before entry into the Land of Israel encompass so many mythical and social insights for the rest of the journey? Dvar Torah on Parshat VaEthanan At the beginning of parashat "Vaethanan" Moses beseeches God: "Let me go over, I pray Thee, and seethe goodland that is beyond theJordan…" God replies to Moses: "Get thee up into the top of Pisgah, and lift up thine eyes westward, and northward, and southward, and eastward, and behold with thine eyes…" It is strange that, when Moses is on the top of the mountain, he is asked to cast his eyes up to see the land, when we would expect that from the top of Pisgah he would look down to see the land which he is prevented from entering. We learn from this that in order to seethe goodland of Israel, we must look up to the heavens. We should rise up from the actual world to see the desired world with the eyes of the spirit. EretzIsrael, the Promised Land, demands the vision in order to be good. In the Book of Proverbs, 29:18, is written: "Where there is no vision, the people cast off restraint." And indeed, inthe lifeofthe community andin our private lives, when there is vision there is meaning – it is the meaning that creates the vision. "Nishma Ve'Na'ase" (We will hear and we will do) When the Torah usesthe root"shema" the meaning is "mashmaut" (meaning) which comes not from the organ of hearing, the ear, but from "an understanding heart" as Solomon requests God to give him: "Give Thy servant therefore an understanding heart to judge Thy people, that I may discern between good and evil; for who is able to judge this Thy great people?" (Kings1 3:9). In the Book of Deuteronomy, the commands begin with "hear" and end with "Do." For example: "And now, O Israel, hearken unto the statutes and unto the ordinances, which I teach you, to do them; that ye may live, and go in and possess the land whichthe LORD, the God of your fathers, giveth you." "And Moses called unto all Israel, and said unto them: Hear, O Israel, the statutes and the ordinances which I speak in your ears this day, that ye may learn them, and observe to do them." "Hear therefore, O Israel, and observe to do it; that it may be well with thee, and that ye may increase mightily," The deed is the tool, the meaning is the tool's purpose. The deed exists only to contain the meaning. A deed with no meaning does not renew anything. It repeats itself, is routine, banal, without imagination, boring! A meaningful deed brings renewal, a new creation, holiness. The Torah, according to the verses cited above, teaches us to understand the meaning before the deed and so we will be blessed with life, success, and fruitfulness. We find four verbs connected to the performance of mitzvoth, in the following order: hear, learn, keep and do. How does each one connect us to mitzvah? "Hearing" is the experience that expands our consciousness and connects us to what is hidden in us and is still waiting to be realized. The children ofIsraelheardthe voiceof God speaking to them fromMount Sinai, awakening their inner voice,the voiceof their souls which had been sleeping when they were slaves of Pharaoh and were only materialistic. They said: "if we will continue to hear God our Lord's voice we will die." The experience of "hearing" is so shattering that a person might repress it because it makes them panic, ignore it as if it wasn't there, deny it. In this situation he must learn. It is not frightening,the personis in control, the intellect reacts, asks and examines, criticizes and acquires knowledge, in the hope of reaching spiritual meaning by experimentation and consideration. "Keeping" is the condition of consciousness, like keeping in a safe, like keeping a secret. The experience and the learning exist and are known in man's consciousness, what remains now is action in the full meaning of the word. Moses, this great man, who brought us fromEgypt, brought us to theJordanand completed his role. His entreaties (Meaning: Grant me from your grace, not because I deserve it) did not help him. He was the process. His book – the Book of Deuteronomy – perpetuates the process as it is: Hear and act. As if he is telling his people before entering the land to build a community there: When " action" precedes "hearing" you have made a golden calf. Inthe lifeof "hearing and doing": ye may live, and that it may be well with you, and that ye may prolong your days in the land which ye shall possess" Shabbat Shalon U'Mevorach
1.456146
HuggingFaceFW/fineweb-edu
I recently noticed that both my kids were becoming sensitive to having dirty hands. The two kids who pick up spiders, worms, play in dirt and sand all summer long don't like to have wet "dirty" hands. During a recent activity with paint they were very uncomfortable with the thought of having paint on their hands and both requested a baby wipe to keep near by. I can respect that they want to stay neat and clean but this may some day become a problem if they want to become chefs, sculptors or painters. So, I conducted an intervention of sorts and we called it GOOP week. I made a batch of "flubber" that I found a recipe for on this website. I have since found that Borax was probably not safe for us to play with, but of course it was fun at the time. (Live, learn, and adjust.) J & A were both pretty apprehensive about getting their hands dirty at first but with some coaxing they did great. How funny it is typing that I had to encourage my toddlers to get dirty. The second goo I made with them was a mixture of corn starch, baking soda, and water. We added food coloring for a "wow" factor which of course the kids thought was cool. I gave them funnels, bowls, and cups to use with their goos. They found out quickly that the flubber didn't go through the funnel. Mixing with their hands. I did notice that they both liked the two different goos but they played with the corn starch mixture for a little while longer. I think this was because they were able to transfer the liquid from one container to the other with greater ease. Another activity for goop week was sort of a free for all. I gave them bowls that contained some ingredients that they could mix. There was flour, rice, oatmeal, lentils, and water. I provided them with measuring spoons, measuring cups, and mixing bowls. Before A & J began mixing, we took the time to discuss texture. They talked about how things felt when they touched them. This looked like a lot of gooey and sticky fun. It wasn't long before J asked me if he could use a muffin pan. I feel that this goop activity really pushed A & J. It required them to work with an extremely sticky substance and it definitely led to wet dirty hands. It was interesting to note that even when given he exact same ingredients they used them in varying combinations. A's goo was the consistency of dough whereas J's goo had the consistency of batter. It is was interesting side note.
1.072012
HuggingFaceFW/fineweb-edu
In this post I'll describe a computational method for splitting two sides of a cell biological structure. It's a simple method that relies on principal component analysis, otherwise known as PCA. Like all things mathematical there are some great resources on the web, if you want to understand this operation in more detail (for example, this great post by Lior Pachter). PCA can applied to many biological problems, you've probably seen it used to find patterns in large data sets, e.g. from proteomic studies. It can also be useful for analysing microscopy data. Since our analysis using this method is unlikely to make it into print any time soon, I thought I'd put it up on Quantixed. During mitosis, a cell forms a mitotic spindle to share copied chromosomes equally to the two new cells. Our lab is working on how this process works and how it goes wrong in cancer. The chromosomes attach to the spindle via kinetochores and during prometaphase they are moved to the middle of the cell. Here, the chromosomes are organised into a disc-like structure called the metaphase plate. The disc is thin in the direction of the spindle axis, but much larger in width and height. To examine the spatial distribution of kinetochores on the plate we wanted a way to approximately separate kinetochores on one side if the plate from the other. Kinetochores can be easily detected in 3D confocal images of mitotic cells by particle analysis. Kinetochores are easily stained and appear as bright spots that a computer can pick out (we use Imaris for this). The cartesian coordinates of each detected kinetochore were saved as csv and fed into IgorPro. A procedure could then be run which works in three steps. The code is shown at the bottom, it is wrapped in further code that deals with multiple datasets from many cells/experiments etc. The three steps are: - Analysis on each subset I'll describe each step and how it works. 1. Principal component analysis This is used to find the 3rd eigenvector, which can be used to define a plane passing through the centre of the plate. This plane is used for division. Now, because the metaphase plate is a disc it has three dimensions, the third of which – "thickness" – is the smallest. PCA will find the principal component, i.e. the direction in which there is most variance. Orthogonal to that is the second biggest variance and orthogonal to that direction is the smallest. These directions are called eigenvectors and their magnitude is the eigenvalue. As there are three dimensions to the data we can get all three eigenvectors out and the 3rd eigenvector corresponds to thickness of the metaphase plate. Metaphase plates in cells grown on coverslips are orientated similarly, but the cells themselves are at random orientations. PCA takes no notice of this and can simply reveal the direction of the smallest dimension of a 3D structure. The movie shows this in action for a simulated data set. The black spots are arranged in a disk shape about the origin. They are rotated about x by 45° (the blue spots). We then run PCA and show the eigenvectors as unit vectors (red lines). The 3rd eigenvector is normal to the plane of division, i.e. the 1st and 2nd eigenvectors lie on the plane of division. Also, the centroid needs to be defined. This is simply the cartesian coordinates for the average of each dimension. It is sometimes referred to as the mean vector. In the example this was the origin, in reality this will depend on the position and the overall height of the cell. A much longer method to get the eigenvectors is to define the variance-covariance matrix (sometimes called the dispersion matrix) for each dimension, for all kinetochores and then do an eigenvector decomposition on the matrix. PCA is one command, whereas the matrix calculation would be an extra loop followed by an additional command. The distance of each kinetochore to the plane that we defined is calculated. If it is a positive value then the kinetochore lies on the same side as the normal vector (defined above). If it is negative then it is on the other side. The maths behind how to do this are in section 10.3.1 of Geometric Tools for Computer Graphics by Schneider & Eberly (starting on p. 374). Google it, there is a PDF version on the web. I'll save you some time, you just need one equation that defines a plane, Where the unit normal vector is [a b c] and a point on the plane is [x y z]. We'll use the coordinates of the centroid as a point on the plane to find d. Now that we know this, we can use a similar equation to find the distance of any point to the plane, Results for each kinetochore are used to sort each side of the plane into separate waves for further calculation. In the movie below, the red dots and blue dots show the positions of the kinetochores on either side of the division plane. It's a bit of an optical illusion, but the cube is turning in a right hand fashion. 3. Analysis on each subset Now that the data have been sorted, separate calculations can be carried out on each. In the example, we were interested in how the kinetochores were organised spatially and so we looked at the distance to nearest neighbour. This is done by finding the Euclidean distance from each kinetochore to every other kinetochore and putting the lowest value for each kinetochore into a new wave. However, this calculation can be anything you want. If there are further waves that specify other properties of the kinetochores, e.g. brightness, then these can be similarly processed here. The code in its present form (not very streamlined) was fast and could be run on every cell from a number of experiments, reading out positional data for 10,000 kinetochores in ~2 s. For QC it is possible to display the two separated coordinated sets to check that the division worked fine (see above). The power of this method is that it doesn't rely on imaging spindle poles or anything else to work out the orientation of the metaphase plate. It works well for metaphase cells, but cells with any misaligned chromosomes ruin the calculation. It is possible to remove these and still fit the plane, but for our analysis we focused on cells at metaphase with a defined plate. What else can it be used for? Other structures in the cell can be segregated in a similar way. For example, the Golgi apparatus has a trans and a cis side, which could be similarly divided (although using the 2nd eigenvector as normal to the plane, rather than the 3rd). Acknowledgements: I'd like to thank A.G. at WaveMetrics Inc. for encouraging me to try PCA rather than my dispersion matrix approach. The post title comes from "Division Day" a great single by Elliott Smith.
2.388383
openbmb/Ultra-FineWeb
necessary to translate the target with this arrangement, not rotate and translate. However, it is still possible with this arrangement to rotate a simple disk target about an axis normal to its face if desired. The translate only feature is needed by the apparatus of FIG. 2 where the target 20 is one of many cylindrical segments on a rotating drum 22. In this case, the rotation axis is perpendicular to the plane of the incident laser beams, so that different target materials can be rapidly selected. Scanning of the beam across the target segment is accomplished by dithering the firing time of the laser relative to the drum rotation angle which is equivalent to simple translation. The second advantage of using the divided beam arrangement is to take a typical gaussian spatial intensity profile from the laser which is most intense at the center, split it upper and lower, and then recombine the two halves at the target such that the less intense edges are overlapped by the more intense centers. As seen in FIG. 3, when the beam 14 is split, a cooler (less intense) top edge 30 of the beam 14 is located in the upper half 14a and the cooler bottom edge 32 is located in the lower half 14b. The hotter middle portion of the beam is located in both halves along edges 34 and 36. This overlapping low and high intensity areas of the laser beam smooths out the intensity profile giving much more uniform illumination of the ablation spot. This helps to keep the ablation plume composition more uniform and results in more uniform films. The same principle can be applied to the left-right portions of the laser beam if desired, but here one would accomplish the effect with a very shallow prism upstream of the lenses. Typical excimer lasers do not suffer as much beam profile variation left to right as up and down. This technique can greatly reduce if not eliminate the need to mask off and discard the "non-uniform" portions of the laser beam as is currently practiced in laser ablation. The two beam halves 14a and 14b pass through two quartz vacuum windows 38 and 40 to inter into the vacuum chamber 42. While the reflecting beam divider 12 is shown as a solid piece, it can be accomplished with two separate mirrors. Also, the embodiment of FIG. 2 shows the use of two windows 38 and 40, but this could be reduced to one by bringing the mirrors 24 and 26 and beam divider 12 inside the vacuum chamber 42. Since mirrors are used to direct the beam, it could also be brought in from the left as well, with the mirrors inside. Ablated material forms a plume 43 which extends substantially normal to a substrate 45 mounted on a support 47. Window fogging protection means is provided by two quartz plates mounted inside the ablation vacuum chamber 42 so as to intercept any plume material that would otherwise strike the vacuum windows 38 and 40 and result in fogging and reduced energy transmission. The angle that the laser beam goes through the plates is chosen to be Brewster's angle which reduces the reflection loss at the plate to near zero for a polarized laser beam. Thin plates are preferred for cost, mounting, handling, and optical absorption minimization. Also, the plates 44 and 46 do not form part of the pressure boundary of the chamber 42, as do the windows 38 and 40, and can thus be thinner. Mask plates 48 and 50 are mounted inside the quartz plates 44 and 46 to shield the unused areas of the quartz plates from the ablated material. The mask plates 48 and 50 can be any suitable metal or other vacuum compatible material. Each has a hole 52 and 54, respectively, sized to permit passage by respective beam halves. When fogging of the beam area of the quartz plate occurs, a fresh area can be moved into place. For example, the quartz plates 44 and 46 can be disk-shaped, and rotated by shafts 56 and 58 connected to suitable actuators (not shown). The shafts 56 and 58 pass through seals 60 and 62, respectively, mounted in opposite sides of the chamber 42. As a result, a fresh area of the plates 48 and 50 can be selected without breaking the vacuum. Since the plates can be much larger than the laser spot size, many fresh areas are available, permitting much longer operation of the ablation process before needing to replace the quartz plates 48 and 50. When replacement is needed, the plates are removed and cleaned in an acid solution in the case of high temperature superconductor materials. The plates are inexpensive compared to vacuum windows and could even be discarded. Other arrangements of plates and actuators are possible. The plates could be rectangular with a translator to move them in two dimensions for maximum area usage. The actuator could be inside the vacuum chamber and electrically powered. An on-line quartz plate transmission monitor can be used to determine when it is time to move the plates. As shown in FIG. 2, the monitor includes a light source 64 and photodetector 66. The arrangement permits the simultaneous transmission detection of both plates 44 and 46. The monitor could be connected to automatically index the plates to fresh areas when the transmission falls below a preset value, allowing very long uninterrupted ablation runs. The transmission monitor can use a simple laser if desired, or even an incoherent light source. The source and photodetector could be mounted outside the vacuum chamber if windows are provided. The invention can be used with other conventional devices such as optical plume monitors, ion probe plume monitors, and heater/substrate holders. It is compatible with on-line optical and ion-probe plume monitoring. Simple plume monitoring techniques usually incorporate either a fixed optical beam through the chamber or a fixed ion probe. In order to compare successive plumes, probing is done in the same spatial part of the plume relative to the ablation spot on the target. If the spot is being scanned up and down or back and forth as is reported by some researchers as described hereinabove, then the monitors also have to be scanned accordingly. The use of dual beams solves the texturing problem with target motion alone, leaving the ablation spot fixed relative to the chamber. Conceivably, the beam dividing and converging functions could be combined into a single optical device capable of directing two halves of the input beam onto the surface of the target. The present invention results in more efficient use of the laser light and will give similar performance from a significantly less expensive laser than the conventional arrangement, because it does not need to aperture the laser in the manner used by most of the workers in the field of laser ablation today. The invention eliminates the texturing from the start, keeping the ablation process much more stable with target consumption. The present invention also lends itself to the use of continuously fed targets such as a ribbon or rod. It works with a line focus which is optimum for most high power excimer lasers which have rectangular beam apertures and different horizontal and vertical beam divergences. Preferably, the beams are brought in symmetrically so that the plume interaction with the beam is uniform across the width of the plume and across the width of the beam. Also, while the embodiment described herein uses a multi-segment drum target, it will be effective with a conventional disk target rotated about its axis with the ablation spot fixed (relative to the chamber). Higher and more uniform usage of such a target could be obtained if such a target's motion were compound as is typically used in an orbital sander or buffer. Here again, though, the laser beams and ablation spot would remain stationary. If one wanted to deposit films over large area substrates, the substrate would be moved back and forth or rotated and translated to expose the larger area to the fixed plume. The present invention will be useful in laser ablation materials preparation or investigation research programs throughout government, private industry and university laboratories. It will be particularly valuable to any industrial manufacturing applications of laser ablation. It is anticipated to be useful in ablation applications where the object is simply to remove material, leaving a smooth surface as is the case in laser paint stripping, for example. It is also expected to be useful in the construction of reliable, long running non-equilibrium materials deposition systems, and will be useful in materials analysis equipment where laser ablation is used to sample the surface being analyzed. The aforementioned uses are by way of example, and not intended to be limiting. Applications of the present invention go beyond mere film deposition. For example, laser eye surgery devices can use the dual beam concept to ensure a more smooth finished surface. While advantageous embodiments have been chosen to illustrate the invention, it will be understood by those skilled in the art that various changes and modifications can be made therein without departing from the scope of the invention as defined in the appended claims. |Cited Patent||Filing date||Publication date||Applicant||Title| |US4128752 *||Dec 15, 1976||Dec 5, 1978||Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of National Defence||Laser micromachining apparatus| |US4315130 *||Feb 28, 1979||Feb 9, 1982||Hitachi, Ltd.||Method of treating object by laser beam and apparatus therefor| |US4623776 *||Jan 3, 1985||Nov 18, 1986||Dow Corning Corporation||Ring of light laser optics system| |US4645900 *||Dec 10, 1985||Feb 24, 1987||Winkler & Dunnebier, Maschinenfabrik und Eisengiesserei GmbH & Co. KG||Apparatus for cutting a moving material web to shape by burning with a laser beam| |US50
1.557858
m-a-p/FineFineWeb
Putting Names in Alphabetical Order in Microsoft Excel by Nick Peers If you have a list of names in your Excel spreadsheet, you can put the names in alphabetical order by using the Sort feature. You can sort the list in ascending or descending order. To maintain the integrity of your data, you must sort all columns in the spreadsheet at the same time when you sort the list of names. You can also sort only the list without changing the order of the values in the other columns. 1 Open the spreadsheet you want to edit in Excel. To open a spreadsheet, press "Ctrl-O." 2 Select the entire spreadsheet if you want to maintain the integrity of the data. When you sort the list of names, all other columns will be sorted automatically. 3 Click the "Data" tab and then click the "Sort" button in the "Sort & Filter" group to open the Sort window. 4 Check the "My data has headers" if your spreadsheet has headers. Headers will not be sorted. 5 Select the column that contains the names from the "Sort By" drop-down box in the Column field. 6 Select the order from the "Order" drop-down box. The "A to Z" option puts the names in ascending order, while the "Z to A" option puts them in descending order. 7 Click "OK" to sort the names and put them in alphabetical order. Verify the data and then press "Ctrl-S" to save the changes. Tip - To sort a single column and leave the rest of the spreadsheet intact, select the column that contains the names by clicking its header and then click the "Sort A to Z" or "Sort Z to A" button in the "Sort & Filter" group. Click "Continue with the current selection" and then "Sort." Warning - Information in this article applies to Microsoft Office Excel 2013. It may vary slightly or significantly with other versions or products. Video of the Day Brought to you by Techwalla Brought to you by Techwalla About the Author Nick Peers has been writing technology-related articles since 2003. His articles have appeared in dozens of technical publications, including MSN UK, CNET, BBC Who Do You Think You Are, LifeHacker UK and TechRadar. He holds a Masters in information technology degree from the University of East London. More Articles Photo Credits
1.747862
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
Beyond the Once the cattle are relaxed and ready, it's time to leave the farm. During transfer, every possible precaution is taken to ensure that the animals are well treated and remain calm and comfortable at all times. We make it clear that ensuring animal safety is paramount at every stage of the journey. The animals are tracked using a passport system to guarantee full traceability, and when they reach their destination they are bedded down for the night in comfortable farm-like surroundings. We consistently maintain the highest standards and go to great lengths to ensure that all our animals are kept calm and treated with respect at all times.
1.203647
Zyphra/Zyda-2
Henderson Fire Stop the Bleed No matter how rapid the arrival of professional emergency responders, bystanders will always be first on the scene. A person who is bleeding can die from blood loss within five minutes, therefore it is important to quickly stop the blood loss. Remember to be aware of your surroundings and move yourself and the injured person to safety, if necessary. Always call 911 in an emergency first. You may be able to save a life by taking simple actions immediately. Here are three steps you can take to help save a life: If you do have a trauma first aid kit: For life-threatening bleeding from an arm or leg and a tourniquet is NOT available OR for bleeding from the neck, shoulder or groin: 1. Open the clothing over the bleeding wound. 2. Wipe away any pooled blood. 3. Pack (stuff) the wound with bleeding control gauze (preferred), plain gauze, or clean cloth. 4. Apply steady pressure with both hands directly on top of the bleeding wound. 5. Push down as hard as you can. 6. Hold pressure to stop bleeding. Continue pressure until relieved by medical responders. If you do not have a first aid kit: 1. Take any clean cloth and cover the wound. 2. If the wound is large and deep, try to "stuff" the cloth down into the wound. 3. Apply continuous pressure with both hands directly on top of the bleeding wound. 4. Push down as hard as you can. 5. Hold pressure to stop bleeding. Continue pressure until relieved by medical responders. Apply the tourniquet 1. Wrap the tourniquet around the bleeding arm or leg about 2 to 3 inches above the bleeding site (be sure NOT to place the tourniquet onto a joint-go above the joint if necessary) 2. Pull the free end of the tourniquet to make it as tight as possible and secure the free end. 3. Twist or wind the windlass until bleeding stops. 4. Secure the windlass to keep the tourniquet tight. 5. Note the time the tourniquet was applied. Note: A tourniquet will cause pain but it is necessary to stop life threatening bleeding.
1.577787
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
-defined. The available data are too sparse to determine if subjective responses can be caused by RF exposure, although the strongest studies indicate no effects on a range of endpoints. Disruption of complex behavioral performance in several animal species, under diverse exposure conditions, has been used as a basis for setting human exposure guidelines since 1982. The threshold SAR selected to establish the standard was chosen at 4 W/kg, a level based upon thermal effects and often (but not always) accompanied by an increase in core body temperature of ~1.0°C. Alteration of an assortment of other behaviors, both learned and unlearned, can also occur in animals at SARs between 1and 4 W/kg, subject to the frequency of the signal and the size of the animal. It appears that the behavioral changes due to RF exposure at these levels are reversible, and no consistent evidence exists for long-term, permanent effects. Extrapolation of animal data to humans has been useful in setting exposure standards. However, human ability to discriminate and cognitively act upon perception of intense RF fields is generally superior to that ability in animals and therefore animal data may tend to underestimate threshold levels for safety. OTHER PHYSIOLOGICAL STUDIES EEG and Brain Electrical Activity Humans exposed to mobile-phone RF fields have generally not shown effects of exposure on the spontaneous, awake electroencephalogram (EEG) (Hietanen and others 2000; Roschke and Mann 1997). However, possible changes have been observed when exposure was given during the performance of memory tasks or under other more demanding paradigms (Freude and others 1998, 2000; Eulitz and others 1998. Krause and others reported changes in EEGs of humans exposed to 902 MHz for both auditory tasks (Krause and others OCR for page 117 An Assessment of Potential Health Effects from Exposure to Pave Paws Low-Level Phased-Array Radiofrequency Energy 2000a) and visual memory tasks (Krause and others 2000b). Additionally, Croft and others (2002) determined the resting EEG parameters to be changed during an auditory discrimination task. Several studies evaluating sleep parameters have examined EEG in human subjects as a way to assess RF impacts on sleep. These studies have been discussed above in the section on behavior and RF exposure in humans. Studies in a variety of animals, including rats, rabbits, cats, and monkeys have shown various changes in EEG measurements from the brain (rabbits at 30 MHz, 0.5-2 kV/m, Takashima and others 1979; cats at 147 MHz, unspecified SAR, Bawin and others 1973; rats at 945 MHz, unspecified SAR, Vorbyov and others 1997; rats at 900 MHz, 1.3 W/kg, Thuroczy and others 1994); however the types of changes are not consistent across studies nor have they generally been independently and systematically confirmed. Effects on Blood Pressure/Heart Rate A stimulation study in humans by Braune and others (1998) initially reported increases in heart rate (HR) and blood pressure (BP). However, these effects were not replicated in the same laboratory (Braune and others 2002), and were also not confirmed by an additional independent human study (Tahvanainen and others 2004). Animal studies have reported effects of RF exposure on BP and HR and other cardiac functions (e.g., Lu and others 1992); however, these have all been conducted at exposure levels in which thermal increases would be expected in the tissue. Blood Brain Barrier Studies Using RF exposures of 2450 MHz, Frey and others (1975) initially reported that at approximately 1 W/kg an increase in blood brain barrier (BBB) permeability was observed in rats. Using a 1.3 GHz RF signal, Oscar and Hawkins (1977) reported increased BBB permeability at 0.4 W/kg (CW) and 0.1 W/kg (PW). Preston and others (1979) suggested that the changes observed by Oscar and Hawkins may have been due to variations in blood flow, so Oscar and others (1981) subsequently determined that increased local brain blood flow did occur following RF exposure. In following up this finding, Oscar and co-workers used a technique to measure BBB permeability that is insensitive to blood-flow change, and reported no effect of RF radiation on BBB (Gruenau and others 1982). Thus the effect originally reported by Oscar and Hawkins (1977) was probably an artifact. Using techniques similar to those of Oscar and Hawkins, at 2450 MHz, Preston and others (1979) and Preston and Prefontaine (1980) reported no effect of RF exposure on BBB permeability at whole body SARs (0.02-6 W/kg) or at SARs in the head (0.08-1.8 W/kg). Additional efforts to establish the occurrence of the BBB effects observed by Oscar and Hawkins (1977) and Frey and others OCR for page 118 An Assessment of Potential Health Effects from Exposure to Pave Paws Low-Level Phased-Array Radiofrequency Energy (1975) have been unsuccessful (Ward and others 1982; Ward and Ali 1985; Merritt and others 1978). Pigs have been exposed repeatedly to 452 MHz fields intermittently for 8 h/d for 90 days. The BBB showed no leakage in exposed animals, nor did neurohistological and enzyme-histochemical preparations show any evidence of damage to nervous tissue in the brain (Sutton and others 1982). Further animal studies in mice also have demonstrated no BBB permeability changes with either a one-hour exposure at 4 W/kg (whole body) or after a lifetime of exposure at SARs ranging from 0.25, 0.5, 1.0, and 4.0 W/kg (whole body) (Finnie and others 2001, 2002). At 2450 MHz, Sutton and Carroll (1979) observed that BBB permeability was increased in rats when the temperature in the brain was 40° C or more. Furthermore, when the core body temperature of the rat was kept at 30° C during exposure of the head, the exposure time had to be lengthened to produce any disruptive effects on the BBB. These observations suggest that RF-induced hyperthermia may indeed be the cause of BBB disruption during exposure. Merritt and others (1978) also showed that BBB permeation in rats was impacted by providing either hot air or RF radiation to heat the animals to 40°C and concluded that hyperthermia was the causative factor, not RF energy per se. Williams and colleagues (1984a-d) report on a series of experiments in which they conclude that RF exposure (at 2450 MHz) produced BBB effects that result from temperature-dependent changes and not as a direct result of the RF energy. Fritze and others (1997) also found BBB permeability changes in rats consistent with thermal effects. A number of other papers have also demonstrated changes in BBB permeability resulting from thermal effects of RF exposure (Lin and Lin 1980, 1982; Goldman and others 1984; Neilly and Lin 1986; Moriyama and others 1991; Ohmoto and others 1996). There are a few papers that report BBB-permeation effects in animals exposed to RF fields below those considered to be "thermal." Persson and others (1997) observed an increase in BBB permeability by about three-fold in animals exposed to CW 915 MHz radiation. However, the changes did not vary with SAR from 0.02 to 8.3 W/kg. Results using modulated RF exposure were not SAR dependent either, with the lowest SARs (0.0004-0.008 W/kg) demonstrating the greatest permeability changes, and at the highest SARs (1.7-8.3 W/kg) no modulated frequency was effective in increasing BBB permeation. The 1997 paper by Persson and others appears to include data from previous studies in their laboratory (Salford and others 1993, 1994; Persson and others 1992) and a recent paper (Salford and others 2003) from this research laboratory describes effects of 915 MHz RF on the BBB in rats exposed to very low SARs (< 0.2 W/kg). No exposure-response relationship was found in work performed in another laboratory (Chang and others 1982) in which only one of six RF exposure levels (at 1 GHz) affected BBB permeability in dogs. OCR for page 119 An Assessment of Potential Health Effects from Exposure to Pave Paws Low-Level Phased-Array Radiofrequency Energy IMMUNEAND ENDOCRINE-FUNCTION STUDIES There are very few studies in humans examining endocrine status during or following exposure to RF fields. Mann and others (1998) observed no changes in serum melatonin, growth hormone, or luteinizing hormone levels during exposure of volunteers to 900 MHz, although cortisol production showed a small transient increase. No effect on urinary levels of 6-hydroxymelatonin sulfate was reported in humans exposed to 900 MHz fields (Bortkiewicz and others 2002). In addition, Radon and others (2001) observed no changes in melatonin, cortisol, or markers of immune function when humans were exposed for 4 hours to 900 MHz fields. In animals, most of the studies investigating immune and endocrine function have been performed at 2450 MHz, with only a few
2.503281
HuggingFaceFW/fineweb-edu
Your browser does not support JavaScript Student Funding Board Harvard Law School The task of the Student Funding Board is to distribute funding among student organizations with widely divergent aims and interests. Each year HLS organizations collectively request nearly twice as much funding as is available. This year – with the economy in its current state, and with more student organizations and the same amount in the overall budget – the task will be particularly difficult. While the specific balance of funding priorities changes each year depending upon the membership of the SFB, the key considerations remain largely constant. These guidelines are issued to make the funding process more transparent. The SFB encourages organizations to work together on similar projects and to ensure that a wide range of activities and organizations are funded; organizations and events that overlap significantly are funded on the assumption that they will co-sponsor some events. The SFB attempts to fund diverse views and student populations to ensure that the HLS is a welcoming and interesting community. Organization size is also a significant consideration for funding determinations to ensure that various populations and interest groups have adequate funds given the size of their constituencies. The SFB believes that open membership policies encourage campus involvement and dialogue; exclusive and limited membership policies are viewed unfavorably. Funding Principles In evaluating allocation requests, the Student Funding Board shall consider the following principles. Although none are requirements, the Board shall look positively upon requesting organizations that advance these principles. Diversity and Inclusion Actively promotes diversity and inclusion on campus and beyond through its activities and policies. Community Service Participates in community service and volunteer work. Skill Acquisition Develops lasting skills for participants and members, including, but not limited to, advocacy, writing, speaking, or negotiating. Ideal skill acquisition involves actively applying skills, rather than passively absorbing knowledge or activities primarily for recreation. Accessibility to Students Proactively works to ensure that its activities are open and accessible to all students, and that they are well publicized to the entire law school community. Responsible Fund Management Has demonstrated the responsible management of its funds from past years, has robust fund management practices in place, and has demonstrated rigorous planning for its activities. Alternative Funding Sources Has sought and received alternative sources of funding to support itself. Alternative sources of funding are looked more favorably upon if they are used to support the organization's activities rather than held in reserve. Has a demonstrable track record of collaboration for its activities in the past, and has concrete plans or secured agreements to collaborate for future activities. Evidence of Effectiveness Collects and implements evidence that its activities are effective. Evidence can include, but is not limited to, levels of attendance, feedback from participants, or frugality of purchases.
1.758237
Zyphra/Zyda-2
Present Tense and Gerund Put the verbs in the correct form. Use only Present Simple and Gerund. Mind the spelling. 1. She (like) (play) soccer with her friends. 2. They (enjoy) (listen) to the radio. 3. We (hate) (play) computer games. 4. He (love) (read) comics 5. (be) you keen on (watch) soap operas? 6. Who (dislike) computer games like Counter Strike? 7. They (dislike) (watch) horror films. 8. She (prefer) (dance) to (sing). 9. Who can't stand (spend) his holidays at the seaside? 10. (do) you really consider (move) to Chicago?
1.286059
Zyphra/Zyda-2
Skip to main content Wind-to-Hydrogen Project Video (Text Version) This text version of the wind-to-hydrogen video describes how electricity from wind turbines is used to produce hydrogen at NREL's National Wind Technology Center. Video Audio 1 Meadow with flowers bobbing in wind Sound of light breeze 2 Tree tops moving with the wind Sound of strong wind 3 Wind turbine spinning Sound of strong wind 4 Darkness Silence 5 Bright light bulb Click of a metal pull-string on light bulb 6 Light bulb hanging from a string goes on; light bulb starts bobbing as wind pushes it around "Here's an idea! Since everybody's talking about renewable energy, why doesn't someone find a way to capture the wind! Yes, we have turbines, but you know...the wind doesn't always blow when we need it the most. Hey, it's just an idea..." 7 Xcel Energy and the National Renewable Energy Laboratory (NREL) "Well, Xcel Energy and the U.S. Department of Energy's National Renewable Energy Lab, or NREL, have teamed up to do just that. They are exploring ways to capture the wind's energy and deliver it more reliably than Mother Nature." 8 Nature showing variable wind "Not that Mother Nature's done anything wrong, but wind is naturally, well, windy." 9 Spinning turbine slowly grinds to a halt "So, while it's one of the least expensive renewable energy sources, it's also rather unreliable." 10 Wind turbines; city lit up at night "Xcel Energy and NREL are testing ways to convert wind energy to hydrogen, which is an energy carrier that can be stored. Then, when energy demand is high, we can use that hydrogen...thereby reducing our need for other energy sources, such as natural gas, oil, and coal." 11 Spinning hydrogen atom "So, how does it work? First, let's talk about hydrogen. Why are we converting wind energy to hydrogen?" 12 Graphic illustration of voice "Hydrogen is the simplest and most abundant element in the universe, but it's almost always found as a compound with other elements. It's an energy carrier, much like electricity, but it's not a primary source of energy, like wind. To use hydrogen, we must extract it from compounds and, until now, that extraction produced greenhouse gasses." 13 Graphic illustration of voice "For example, most hydrogen is 'reformed' from natural gas or other fossil fuels by stripping the hydrogen atoms out, creating greenhouse gas emissions in the process. Or, hydrogen can be produced by splitting water through electrolysis using sources of electricity, such as fossil fuels, which themselves generate air emissions." 14 "Our entire process is green because we are using wind power to run the emissions-free electrolyzers, and then using that green hydrogen to make green electricity when we need it!" 15 Wind turbine, hydrogen, and light bulb "Basically, this project has the potential to use our least expensive renewable energy source (wind) and convert it using a green process (electrolysis) into storable energy (hydrogen) that we can tap on demand!" 16 "Now that you understand why we like hydrogen, let's take a look at some of the main technologies and concepts that we are testing." 17 Wind turbines "There are several steps in this conversion process. We are testing not only the system that's transforming the energy, but also each technical component along the way. Before we discuss the entire process, let's look at some of the individual technologies. We are testing…" 18 Wind turbines "Two variable speed wind turbines that create energy from the wind. One is a 100-kilowatt turbine and the other is a 10-kilowatt turbine." 19 Electrolyzers "Two proton exchange membrane electrolyzers and one alkaline electrolyzer that can produce hydrogen gas from water." 20 Tanks "Metal storage tanks that store hydrogen for later use, and..." 21 Engines "An internal combustion engine that runs on the stored hydrogen, generating electricity that can be sent onto the grid during peak demand." 22 Wind turbine, electrolyzers, hydrogen, storage tanks, engines, and light bulb "Essentially, this project studies how we can capture the energy contained in the wind, convert it into a form that can be stored (hydrogen), and then convert it back to electricity during times when it is needed most to run our homes, our businesses, and someday, our cars!" 23 "Let's go back and look at some of the questions we hope to answer." 24 Wind turbines "We are testing those two turbines, experimenting with the different energies that they produce and how to get that electricity into the electrolyzers, where it can be converted into hydrogen. At the heart of this challenge is the fact that most turbines create AC (or alternating current) electricity, while the electrolyzers require direct current, or DC. Here's what we're doing about that..." 25 100kW wind turbine "The 100-kilowatt turbine. This turbine already has on-board electronics that take the wind's varying AC electricity and turn it into 750- to 800-volt DC. Normally, this DC electricity has to be converted to AC current like we use in our homes...but we're going to skip that step and just use it straight from the turbine, with one exception. In order to be used in our electrolyzers and ultimately converted into hydrogen, we have to step that 800 volts down to 50 to 100 volts DC." 26 10kW wind turbine "The 10-kilowatt turbine. This smaller wind turbine does not have on-board converters, so its 'wild AC' must be converted to DC before it can be turned into hydrogen by the electrolyzers. We designed and built new, integrated wind-to-electrolyzer electronics to get the wind-produced electricity to the electrolyzers more efficiently." 27 Electrolyzers and components "Now we have the wind's energy delivered in a useful form to the electrolyzer. An electrolyzer has one main goal: to take in water and electricity and produce hydrogen and oxygen." 28 Proton Energy Systems HOGEN 40RE proton exchange membrane (PEM) electrolyzers and Teledyne HMXT-100 alkaline electrolyzer "We are testing two types of electrolyzers to compare their performance: two HOGEN 40RE proton exchange membrane electrolyzers from Proton Energy Systems and one Teledyne HMXT-100 alkaline electrolyzer." 29 Hydrogen coming from electrolyzers "Now, the hydrogen from the electrolyzers must be compressed so that it can be stored in tanks." 30 Compressor and components "Like air compressors found in mechanics' workshops, our system compresses the hydrogen gas that's coming out of the electrolyzers from 150 pounds per square inch to a maximum pressure of 3,500 pounds per square inch. The higher pressure allows more hydrogen to be stored in the tanks." 31 Hydrogen moving from compressor to tanks and then from tanks to engine or filling station "Once the hydrogen is stored, it can be sent to an internal combustion engine for use on the grid when it is needed or, someday, to a hydrogen automotive filling station." 32 Combustion engine/grid and components "The stored hydrogen will run an internal combustion engine, generating electricity that we'll send onto the grid. We'll run the engine between the hours of 4 and 7 pm, coinciding with one of Xcel Energy's daily peak energy periods. This step can also be accomplished by using a hydrogen fuel cell." 33 Wind in meadow, wind turbine, internal combustion engine, light bulb "Thus, we have taken the variable nature of wind, captured it, and used it when our customers need it most." 34 Hydrogen filling station/cars and components "The hydrogen can be used today to power internal combustion engine vehicles that run on hydrogen gas and eventually to power fuel cell vehicles, which will start coming on the market in the next decade. Imagine, cars with no greenhouse gas emissions, only water." 35 "Of course, we are very excited about the possibilities offered by this study." 36 Isolating hydrogen without creating greenhouse gasses; creating several uses for and improvements on wind and hydrogen energies; comparing electrolyzer technologies; reducing energy loss between wind turbines and electrolyzers; exploring a way to provide consistent support of the grid using a renewable energy "Xcel Energy and NREL are...learning how to isolate hydrogen without creating greenhouse gasses, creating several uses for and improvements on wind and hydrogen energies, comparing electrolyzer technologies, reducing energy loss between wind turbines and electrolyzers, and exploring a way to provide consistent support of the grid using renewable energy." 37 "We hope to have these systems running and providing us with new data throughout 2007 and 2008." 38 Music transition to indicate the show is closing. "Hydrogen technologies have the potential to reduce our nation's dependence on foreign oil, improve air quality, and help the United States maintain its economic competitiveness." 39 Wind, wind turbine, electrolyzers, hydrogen, storage tanks, and engines fade to black "With this project, the National Renewable Energy Laboratory and Xcel Energy are exhibiting leadership and a unique spirit of teamwork to explore a potentially practical, safe solution to meet our country's energy needs." 40 Light bulb turns on
1.723844
Zyphra/Zyda-2
Death is the permanent end of the life of a biological organism. Death may refer to the end of life as either an event or condition. Many factors can cause or contribute to an organism's death, including predation, disease, habitat destruction, senescence, malnutrition and accidents. The principal causes of death in developed countries are diseases related to aging. In medicine, biological details and definitions of death have become increasingly complicated as technology advances. Traditions and beliefs related to death are an important part of human culture, and central to many religions. Nearly every culture has had some belief in an afterlife, some form of continued existence after the death of the human body. While science has lacked the technology to find scientific evidence of such existence, faith and personal experience provide support and comfort for many who understand the purpose of our life to transcend our bodies. Whatever one's belief about death, as it draws near, its finality provides a perspective upon one's entire lifetime. For many people, how one dies is as important as how one lived. Hence people regard it of great importance to take care of unfinished matters and reconcile with loved ones in preparation for death. Historically, attempts to define the exact moment of death have been problematic. Death was once defined as the cessation of heartbeat (cardiac arrest) and of breathing, but the development of CPR and prompt defibrillation have rendered the previous definition inadequate because breathing and heartbeat can sometimes be restarted. This is now called "clinical death." Events which were causally linked to death in the past no longer kill in all circumstances; without a functioning heart or lungs, life can sometimes be sustained with a combination of life support devices, organ transplants, and artificial pacemakers. Today, where a definition of the moment of death is required, doctors and coroners usually turn to "brain death" or "biological death": People are considered dead when the electrical activity in their brain ceases (unlike a persistent vegetative state). It is presumed that a stoppage of electrical activity indicates the end of consciousness. However, suspension of consciousness must be permanent, and not transient, as occurs during sleep, and especially a coma. In the case of sleep, EEGs can easily reveal the difference. Identifying the moment of death is important in cases of transplantation, as organs for transplant must be harvested as quickly as possible after the death of the body. The possession of brain activity, or ability to resume brain activity, is a necessary condition to legal personhood in the United States: "It appears that once brain death has been determined … no criminal or civil liability will result from disconnecting the life-support devices." Those maintaining that only the neo-cortex of the brain is necessary for consciousness sometimes argue that only electrical activity there should be considered when defining death. Eventually it is possible that the criterion for death will be the permanent and irreversible loss of cognitive function, as evidenced by the death of the cerebral cortex. All hope of recovering human thought and personality is then gone. However, at present, in most places the more conservative definition of death—irreversible cessation of electrical activity in the whole brain, as opposed to just in the neo-cortex—has been adopted (for example the Uniform Determination Of Death Act in the United States). Even by whole-brain criteria, the determination of brain death can be complicated. EEGs can detect spurious electrical impulses, while certain drugs, hypoglycemia, hypoxia, or hypothermia can suppress or even stop brain activity on a temporary basis. Because of this, hospitals have protocols for determining brain death involving EEGs at widely separated intervals under defined conditions. Fate of dead organisms After death an organism's remains become part of the biogeochemical cycle. Animals may be consumed by a predator or scavenger. Organic material may then be further decomposed by detritivores, organisms which recycle detritus, returning it to the environment for reuse in the food chain. Examples include earthworms, woodlice and dung beetles. Microorganisms also play a vital role, raising the temperature of the decomposing material as they break it down into simpler molecules. Not all material need be decomposed fully however; for example coal is a fossil fuel formed in swamp ecosystems where plant remains were saved by water and mud from oxidization and biodegradation. In animals, small movements of the limbs (for example twitching legs or wings) known as a postmortem spasm can sometimes be observed following death. Pallor mortis is a postmortem paleness which accompanies death due to a lack of capillary circulation throughout the body. Algor mortis describes the predictable decline in body temperature until ambient temperature is reached. Within a few hours of death rigor mortis is observed with a chemical change in the muscles, causing the limbs of the corpse to become stiff (Latin rigor) and difficult to move or manipulate. Assuming mild temperatures, full rigor occurs at about 12 hours, eventually subsiding to relaxation at about 36 hours. Decomposition is not always a slow process however—for example fire is the primary mode of decomposition in most grassland ecosystems. Some organisms have hard parts such as shells or bones which may not decompose but become fossilized. Fossils are the mineralized or otherwise preserved remains or traces (such as footprints) of animals, plants, and other organisms. Fossils vary in size from microscopic, such as single cells, to gigantic, such as dinosaurs. A fossil normally preserves only a portion of the deceased organism, usually that portion that was partially mineralized during life, such as the bones and teeth of vertebrates, or the chitinous exoskeletons of invertebrates. Preservation of soft tissues is extremely rare in the fossil record. Extinction is the cessation of existence of a species or group of taxa, reducing biodiversity. The moment of extinction is generally considered to be the death of the last individual of that species (although the capacity to breed and recover may have been lost before this point). Because a species' potential range may be very large, determining this moment is difficult, and is usually done retrospectively. This difficulty leads to phenomena such as Lazarus taxa, where a species presumed extinct abruptly "re-appears" (typically in the fossil record) after a period of apparent absence. A typical species becomes extinct within ten million years of its first appearance, although some species, called living fossils, survive virtually unchanged for hundreds of millions of years. Only one in a thousand species that have existed remain today. Prior to the dispersion of humans across the earth, extinction generally occurred at a continuous low rate, interspersed with rare mass extinction events. Starting approximately 100,000 years ago, and coinciding with an increase in the numbers and range of humans, species extinctions have increased to a rate unprecedented since the Cretaceous-Tertiary extinction event. This is known as the Holocene extinction event and is at least the sixth such extinction event. Some experts have estimated that up to half of presently existing species may become extinct by 2100. There are many anecdotal references to people being declared dead by physicians and then coming back to life, sometimes days later in their own coffin, or when embalming procedures are just about to begin. Owing to significant scientific advancements in the Victorian era, some people in Britain became obsessively worried about living after being declared dead. A first responder is not authorized to pronounce a patient dead. Some EMT training manuals specifically state that a person is not to be assumed dead unless there are clear and obvious indications that death has occurred. These indications include mortal decapitation, rigor mortis (rigidity of the body), livor mortis (blood pooling in the part of the body at lowest elevation), decomposition, incineration, or other bodily damage that is clearly inconsistent with life. If there is any possibility of life and in the absence of a do not resuscitate (DNR) order, emergency workers are instructed to begin rescue and not end it until a patient has been brought to a hospital to be examined by a physician. This frequently leads to situation of a patient being pronounced dead on arrival (DOA). In cases of electrocution, CPR for an hour or longer can allow stunned nerves to recover, allowing an apparently-dead person to survive. People found unconscious under icy water may survive if their faces are kept continuously cold until they arrive at an emergency room. This "diving response," in which metabolic activity and oxygen requirements are minimal, is something humans share with cetaceans called the mammalian diving reflex. As medical technologies advance, ideas about when death occurs may have to be re-evaluated in light of the ability to restore a person to vitality after longer periods of apparent death. There have been some scientific attempts to bring dead organisms back to life, but with limited success. Causes of human death Death can be caused by disease, accident, homicide, or suicide. The leading cause of death in developing countries is infectious disease. The leading causes of death in developed countries are atherosclerosis (heart disease and stroke), cancer, and other diseases related to obesity and aging. These conditions cause loss of homeostasis, leading to cardiac arrest, causing loss of oxygen and nutrient supply, causing irreversible deterioration of the brain and other tissues. With improved medical capability, dying has become a condition to be managed. Home deaths, once the norm, are now rare in the first world. In third world countries, inferior sanitary conditions and lack of access to medical technology makes death from infectious diseases more common than in developed countries. One such disease is tuberculosis, a bacterial disease which killed 1.7 million people in 2004. Many leading first world causes of death can be postponed by diet and physical activity, but the accelerating incidence of disease with age still imposes limits on human
2.520647
openbmb/Ultra-FineWeb
activator such as comfrey, nettle or yarrow plants, animal or fish material, urine, or old compost, can be placed in the middle of compost heap to start off composting process. DAY 5 – Turn Compost Pile, Let it Sit for a Day - Turn the compost heap over, turning the outside to the inside, and the inside to the outside. To explain how to do this, when turning the compost, move the outside of the pile to a spot next to it, and keep moving material from the outside to the new pile. When the turning is completed, all the material that was inside the pile will be outside and vice versa. - Ensure that moisture stays constant. To test, put gloves on and squeeze a handful of the compost materials, which should only release one drop of water, or almost drips a drop. - On the next day, let the compost pile sit, don't turn it. - TIP: If the compost pile gets too wet, spread it down, or open a hole about 7-10cm (3-4") wide with the handle of the pitchfork, or put sticks underneath for drainage. DAY 7 & DAY 9 – Measure Temperature, Turn Compost, Let it Sit for a Day - Measure the temperature at the core of the compost heap.The compost heap should reach its maximum temperature on these days. As an simple guideline, if a person can put their arm into the compost up to the elbow, then it is not at 50 degrees Celsius, and is not hot enough. Best to use a compost thermometer or a cake thermometer for this purpose.The hot composting process needs to reach an optimum temperature of 55-65 °C (131-149 °F).At temperatures over 65 °C (149 °F), a white "mould" spreads through the compost, which is actually some kind of anaerobic thermophilic composting bacteria, often incorrectly referred to as 'fire blight'. This bacteria appears when the compost gets too hot, over 65 °C and short of oxygen, and it disappears when the temperature drops and aerobic composting bacteria take over once again.Temperature peaks at 6-8 days and gradually cools down by day 18. - Turn the compost heap over every second day (on day 7 and again on day 9). - Allow the compost to rest for on the next day after turning it. - TIP: If the compost pile starts coming down in size quickly, there is too much nitrogen in the compost. - TIP: To heat up the compost faster, a handful of blood & bone fertiliser per pitchfork when turning speeds it up. - TIP: If it gets too hot and smelly and goes down in size, it has too much nitrogen, need to slow it down, throw in a handful of sawdust per pitchfork when turning. DAY 11, 13, 15 and 17 – Turn Compost, Let it Sit for a Day - Continue to turn the compost every 2nd day (on days 11, 13, 15 and again on day 17). - Allow the compost to rest for a day after turning it. DAY 18 – Compost Completed, Ready to Harvest - Harvest completed compost, which will be warm, dark brown, and smell good. - Congratulate yourself for a job well done! - TIP: When the earthworms move into the compost, it's a sign that it's finished and ready, because it's cooled down enough for them and they're in there because it's full of nutrients! Some important points to note: - Locate the compost heap in an area which is protected from too much sun to prevent the compost from drying out, or from heavy rain to avoid water-logging, as both extreme conditions will slow down the composting process. - Space required for for your heap should be about 1.5 x 1.5 metres (5′ x 5′), and enough space in front of it to stand when turning the compost. - Water each layer until it is moist as you build the heap. After three or four days, give the compost air by mixing and turning it over, then turn every two days until the compost is ready, usually in 14-21 days. Remember, frequent turning and aeration is the secret of successful composting. - Turn the compost using a garden fork, or even better, a long-handled pitchfork. - In cold or wet weather, cover the compost heap with a tarp or plastic sheet, to prevent the rain cooling it down, since the water will penetrate into the core of the compost pile. Even though cold outside air will cool the surface, but not the core of the compost heap, by covering it, this prevents some heat loss from the surface to cooler outside air, and retains the heat within the compost heap better. Is My Garden Too Small for Hot Composting? A full–sized hot compost pile can be made successfully in a small courtyard, I know from experience! The first time I tried hot composting was assisting a friend with only a small courtyard in a rental property, who had never tried this process before. For composting materials, he gathered a wheelie bin full of fallen leaves from his local street, a second wheelie bin full of weeds from his garden, and he also purchased a small straw bale for the sake of it. I also helped him collect a few garbage bags of cow manure from an urban farm. It took us under an hour to pile up all the materials in reasonably thin layers of less than 5cm (2″) to build the compost heap. Even though it was his first attempt at hot composting, and in around 18 days, he had over 1 cubic metre of rich, dark, compost to use in his garden. None of the original ingredients could be identified in the final product either, it had a very fine consistency. Best of all, it cost him next to nothing – the straw bale was the only item purchased, and that was more of a gratuitous addition, as the hot compost would have worked just as well without it. Considering that a hot compost pile doesn't really reduce in volume, the biggest issue in small yards and gardens is figuring out what to do with such a large volume of high-quality compost! Ways to Use Compost in the Garden Wondering what to do with over a cubic metre of freshly made compost? - It can be used to improve your soil by digging it through your garden beds. - Don't like digging? Use the compost to start a no-dig garden with the no-dig gardening method, which is my personal preference! - Compost should be always mixed into the soil to improve drainage in heavy clay soils, and to improve water retention in sandy soils when planting new trees. These are just a few ideas to get things started. Happy composting!
2.133797
openbmb/Ultra-FineWeb
Tổng số lượt xem trang Thứ Bảy, 11 tháng 2, 2017 BRAIN TUMOR ELASTiCITY _ INTRAOPERATIVE SWE Abstract Purpose: Objective Shear wave elastography (SWE) enabled living tissue assessment of stiffness. This is routinely used for breast, thyroid and liver diseases, but there is currently no data for the brain. We aim to characterize elasticity of normal brain parenchyma and brain tumors using SWE. Materials and Methods: Patients with scheduled brain tumor removal were included in this study. In addition to standard ultrasonography, intraoperative SWE using an ultrafast ultrasonic device was used to measure the elasticity of each tumor and its surrounding normal brain. Data were collected by an investigator blinded to the diagnosis. Descriptive statistics, box plot analysis as well as intraoperator and interoperator reproducibility analysis were also performed. Results: 63 patients were included and classified into four main types of tumor: meningiomas, low-grade gliomas, high-grade gliomas and metastasis. Young's Modulus measured by SWE has given new insight to differentiate brain tumors: 33.1 ± 5.9 kPa, 23.7 ± 4.9 kPa, 11.4 ± 3.6 kPa and 16.7 ± 2.5 kPa, respectively, for the four subgroups. Normal brain tissue has been characterized by a reproducible mean stiffness of 7.3 ± 2.1 kPa. Moreover, low-grade glioma stiffness is different from high-grade glioma stiffness (p = 0.01) and normal brain stiffness is very different from low-grade gliomas stiffness (p under 0.01). Conclusion: This study demonstrates that there are significant differences in elasticity among the most common types of brain tumors. With intraoperative SWE, neurosurgeons may have innovative information to predict diagnosis and guide their resection. Đăng nhận xét
1.461786
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
Thanksgiving can mean a lot of things, from the usual turkey feast to more idiosyncratic and potentially stressful traditions like a Black Friday campout or a death-defying deep fry. The holiday varies from person to person and family to family, causing real problems for anyone who wants to contemplate the True Meaning of Thanksgiving. Has the national celebration of football and shopping overtaken the giving-thanks part of it all? Is Thanksgiving kind of a downer? How would you even know? For that one day, people were more thankful and felt more positiveLuckily, we don't have to guess. We've got psychology. For answers, we turn to a sub-discipline called positive psychology, which deals primarily in questions of happiness and well-being. It's produced a number of studies on exactly how Thanksgiving makes people feel, and for the most part, the news is good. It's also more nuanced than you might expect: the discipline has largely moved past generalizations like the "holiday blues," which posited a widespread jump in depression between Thanksgiving and New Years. A battery of recent studies have challenged that assumption, tracking well-being on a day-to-day basis and finding fascinating subtleties in the competing feelings associated with the holiday season. One paper published in The Journal of Positive Psychology this February followed 172 college students for the three weeks around Thanksgiving, tracking gratitude, positive affect, and life statisfaction through diary-style entries. Researchers saw a significant bump in both gratitude and positive affect on Thanksgiving: for that one day, people were more thankful and felt more positive overall. Without gratitude, the holiday stress became overwhelming It makes sense, given the long-standing correlation between gratitude and positive affect — but the study didn't stop there. Once researchers controlled for gratitude, imagining a Thanksgiving-like day without the bump in that sentiment, things got messy. Levels of related readings like life satisfaction and "meaning in life" plummeted. Negative affect jumped up. This makes a certain sense too: Thanksgiving is filled with stressful factors, whether it's travel, family conflicts, or high-stakes cooking, any of which could easily chip away at a person's sense of well-being. The result is a kind of Thanksgiving parable in research form. Without all the gratitude in the air, the holiday stress became overwhelming. tweets indicate a jump in overall well-being For other researchers, an easier way to peer into the soul of the holiday is to look at Twitter. This graph from a study published in a journal called Social Networking this April tracks shopping-related tweets during Thanksgiving, with a major Black Friday-related spike visible on the midnight after Thanksgiving itself. Overall, shopping tweets are dwarfed by general gratitude tweets, and more importantly, peak dinner hours see almost no shopping tweets at all, even while global traffic spikes. Shopping may be an important part of the holiday, but if tweets are any indication, people aren't talking about it much during dinner hours. More importantly, tweets indicate a jump in overall well-being. A separate study from Indiana University monitored Twitter for changes in mood during various events like Christmas and Election Day. During Thanksgiving, the researchers saw a huge uptick in the "vigor" metric, which reflects a general sense of energy and enthusiasm. The measure leapt more than four standard deviations above the norm and confirming much of what the earlier diary study suggested. Whatever else is happening, people are excited to sit down and eat. To be fair, these results are based only on self-reported measurements and social media, both flawed metrics when it comes to the complexities of human psychology. Still, the studies do give us a window into the general spirit of the holiday. Despite all the family feuds and travel woes, a sense of gratitude is still at the center of it all, and it does more than you might think to soften the holiday stress. Think of it as a reminder to be thankful, courtesy of your local psychology lab.
1.839957
openbmb/Ultra-FineWeb
Rate this paper - Currently rating - 1 - 2 - 3 - 4 - 5 4.20 / 5 Paper Topic: Humor and Empowerment of Women in Carol Satyamurti's I Shall Paint my Nails Red Carole Satyamurti, a poet and sociologist cleverly weaved the plight of ordinary women in I Shall Paint my Nails Red : women who are seen as a mere background of the societyneglected and lost in the mundane of her existence. While it is easy to discern that I Shall Paint my Nails is nothing but the showcasing of a woman's vanity seeking attention, a closer look at the poem would reveal a deeper and multi-layered analysis of women. In essence, Carole Satyamurti's I Shall Paint my Nails Red was able to show the different roles and interpretations of women of her self, to her daughter, to her lover, the society and to lifethat painting her nails red symbolizes a departure from her customary role in to show an intelligent, interesting and valuable person capable of so much more than her ordinary routines. It is a declaration of a woman's importance The color red is interpreted in poetry as a figure of speech for passion, anger and strong emotions. In this light, Satyamurti used this color to depict a seemingly humorous and trivial poem into one that speaks of a woman's importance. In the first line "because a bit of color is a public service (line 1 " Satyamurti departs from the convention that public service is about service, the economy and production. Essentially, the poem recognizes that it is not only men who are components of production and economics that can perform public service but also women who provide not only entertainment but beauty and compassion. This is affirmed in her second line when Satyamurti declared that "because i am proud of my hands (line 2'connoting that her hands that are used to do household chores are also important. By painting it red, Satyamurti calls attention to the role of women in the society and in the familythat while they are not being paid for work they do everyday of their lives, it should not be neglected. Thus, being proud of one's hands and putting a red paint on it declares that the woman is aware of her economic and sociological role and significance On the other hand, the third and fourth lines of the poem points to the theme of strength and empowerment in women. In the line "because it will remind me I'm a woman (line 3 " and "because I will look like a survivor " the woman is declaring that being a woman means being brave and strong. Thus, by re-affirming her gender identity, the person in the poem is also recognizing that women in general possess strength of characters. Furthermore, because the color red symbolizes intense emotions, it evokes attention thereby, reminding us that women are not ordinary. For instance, in the fifth line "because I can admire them in traffic jams (line 5 " it is easy to note that this is vanity...
1.286749
Zyphra/Zyda-2
July 4, 2012 A recently published study confirms that solar thermal power is largely unrestricted by materials availability. There are, however, some issues that the industry needs to look into soon, like replacing silver in mirrors. In the wake of Chinese export restrictions on rare earth metals, the dependence of some renewable technologies on scarce materials has gained attention. Several players in the wind and PV industry are struggling to get away from excessive use of restricted elements, such as indium or rare earth metals. Meanwhile, there has been a shared notion amongst solar scientists and industry that Concentrating Solar Power (CSP) should "probably" be less restricted, being built mainly on commonplace commodities like steel and glass. A recently published study from Chalmers University of Technology has gone into the details on material issues for CSP. The main conclusion is that CSP does indeed seem to be largely unrestricted, viewing the material requirements compared to the global reserves. In theory, enough solar plants could be built to cover -- at the very least -- five times the current global electricity demand. However, the report also highlights some issues that are likely to pose challenges to the industry. The main point of concern is that silver, today extensively used for reflecting surfaces, will most likely be in short supply in the coming decades even without demand from a booming CSP industry. CSP mirror manufacturers might have to look at other reflective surface materials, such as aluminium, to secure cost competitiveness. "The prospects for strong growth for CSP over the next few decades seem good, but would cause a stir on the global commodity markets," says Dr Erik Pihl, lead author of the scientific article. Following a Greenpeace/IEA SolarPACES/ESTELA growth scenario where CSP reaches 8000 TWh/year in 2050, the solar plants would consume up to 50-120% of today's yearly nitrate salt production, and 5-15% of several common materials such as glass, nickel, magnesium and molybdenum. The report has used data directly from plant manufacturers Cobra and eSolar for trough and tower plants. These plants have somewhat different characteristics when it comes to material use. "Parabolic trough plants tend to use a lot of concrete and iron, while the concept of small heliostat tower plants has a higher use of aluminium and stainless steel," says Erik Pihl. "The common design of a parabolic trough plant also requires more molten salt per MW than a salt-receiver tower plant, even when the former has fewer storage hours. That means that trough plants appear slightly more sensitive than tower plants to possible salt production bottlenecks, unless other storage techniques can be employed." Erik Pihl believes that we can expect to see material demands for plants decreasing as we go for higher steam temperatures and increased plant efficiency. "We see that clearly when comparing a mature design to a novel concept. That does not automatically mean that all material restriction problems will be solved. We might trade a large use of common materials for small quantities of scarce elements. It comes down to what alloys we use in pipes, receivers and turbines." Higher temperatures means more use of high quality steels, but alloy materials such as molybdenum and niobium have restrictions in both stock and production. "There might be enough for CSP alone, but there are many other uses," says Erik Pihl. "That could be a problem in the more distant future. In the short term, substituting silver and increasing nitrate salt production should be the first priority." Other social bookmarking and sharing tools: - Erik Pihl, Duncan Kushnir, Björn Sandén, Filip Johnsson. Material constraints for concentrating solar thermal power. Energy, 2012; DOI: 10.1016/j.energy.2012.04.057 Note: If no author is given, the source is cited instead.
1.928932
HuggingFaceFW/fineweb-edu
Category: What are the Pros and Cons of a Second C-Section? Article Details - Written By: Lainie Petersen - Edited By: Melissa Wiley - Last Modified Date: 19 January 2020 - Copyright Protected: 2003-2020 Conjecture Corporation - Print this Article Free Widgets for your Site/Blog A 2019 study found that drinking 2 or more sodas a day (regular or diet) is linked to a higher risk of early death. more... January 27 , 1973 : The Vietnam War ended. more... A second c-section delivery, also known as a cesarean section, has significant benefits and risks for both mother and baby. A second c-section can be the safest option for a woman who wants to avoid the possibility of her uterine scar rupturing during labor, which can cause severe complications for both the pregnant woman and her baby or for a woman who lives in an area where hospitals are unable to perform very quick crash c-sections. Many women wish to avoid a second c-section for some significant reasons, including faster recovery time, the preservation of the ability to safely carry future pregnancies, and avoidance of another surgery and its attendant risks. Ad The c-section procedure is major abdominal surgery in which a woman's abdomen and uterus are cut open to deliver her baby. While this surgery is commonly performed and has a high success rate, there are risks involved, including those of infection, bad reactions to anesthesia, and problems with healing. In addition, it is stressful for new mothers to recover from surgery while at the same time adjusting to their new baby. Some women also feel cheated by not being able to deliver vaginally and may not be able to breast feed or hold their babies right after they deliver. By delivering a subsequent pregnancy vaginally, sometimes known as a VBAC or vaginal birth after cesarean, a woman can avoid these risks, as well as prevent further scarring to her uterus, which can affect additional pregnancies. Not all women are able to deliver vaginally after a c-section, even if a vaginal birth is planned and attempted. If a woman is carrying a high-risk pregnancy, her baby is very large, or the baby is in any other position than head down, a second c-section may be the safest option. If a pregnant woman has a scar from a previous c-section on a part of her uterus that is thin and vulnerable to rupture, a second c-section may also not be in her best interests. Women and their doctors should also consider the capabilities of their local medical facilities. While some hospitals are well prepared to intervene with a c-section during a troubled delivery, others may not be able to respond quickly enough in a situation where a woman's previous c-section scar ruptures or another, unexpected complication occurs. If complications do occur under such circumstances, any delay could be deadly to mother, child, or both. Ad You might also Like Recommended Discuss this Article Post your comments Post Anonymously Login username password forgot password? Register username password confirm email
1.329115
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
Jan. 27, 2012 Dr. Susan Mallery, a professor in the College of Dentistry at The Ohio State University and Oral Pathology Consultant at the Ohio State University and James Cancer hospitals, has dedicated her nearly 30-year career to studying new strategies to preventing oral cancer. Oral cancer is currently responsible for more than 7,000 deaths each year, and has a particularly high mortality rate. Treatment relies on excising cells before they turn cancerous, but as many as one-third of all patients will experience a recurrence within a year. "While not all oral lesions progress to cancer, we cannot accurately predict which will be the 'bad actors.' This often results in multiple surgeries and high anxiety in both our patients and clinicians," says Mallery. Since 2003, Mallery has been investigating a variety of agents ranging from anti-angiogenesis drugs to natural products, to identify new therapeutics that can suppress the conversion of pre-cancerous to cancerous cells. Her first breakthrough was the creation of an oral gel based on anthocyanins, powerful antioxidants found in black raspberries. Study results showed that the gel, when applied to the mouth, would suppress genes associated with functions that allow cancerous cells to grow, thus diminishing the risk for recurring lesions. Mallery next turned her attention to identifying alternatives to the surgical removal of pre-cancerous lesions. Mallery thought she might have good possibilities with fenretinide, a decades-old breast cancer treatment that had been largely abandoned by oral pathologists despite strong clinical evidence that showed fenretinide can induce cell death, encourage the body's natural cellular defense system and strangle blood supply to tumors. "The discrepancy between the established efficacy of fenretinide in other cancer types and in oral cancer was puzzling. Then it occurred to me that maybe the structure of the oral mucosa wasn't allowing a therapeutic amount to reach the lesions, even at high doses," says Mallery. "We knew we needed to come up with a way to deliver the therapy directly to the lesion" In 2009, The Ohio State University College of Dentistry, in collaboration with the Center for Clinical and Translational Science (CCTS) awarded Mallery and her team funding for a project aimed at developing a way to treat precancerous lesions directly in the mouth and preventing recurring lesions. Mallery partnered with two pharmaceutical chemists from the University of Michigan (Drs. Stephen Schwendeman and Kashappa Goud Desai) along with two Ohio State investigators Drs. Gary Stoner and Peter Larsen to develop a first of its kind patch that could stick to the inside of the mouth, and deliver a continuous therapeutic dose of fenretinide directly on the lesion. The research team plans to move the patch into pre-clinical and clinical trials, and is already looking at a combination of fenretinide and anthocyanins, as well as testing a combination therapy using the patch and black raspberry-based gel mouthwash to prevent recurrence of lesions. "If we can effectively treat the lesions with the patch, and then prevent more from coming back, we will completely change -- and improve upon -- the way oral cancer is currently treated," said Mallery. Mallery says the best course of action is to prevent oral cancer in the first place, and provides the following tips: - Use of tobacco products -- particularly burned tobacco -- is the greatest risk factor for development of oral cancer. - Watch what you drink and eat. Alcohol consumption -- in excess and especially during smoking -- can have an additive effect and increase your chances to develop oral cancer. Conversely, a diet high in fruits and vegetables can lower your risk. - Infection with human papilloma viruses (HPV) causes a subset of oral cancers. Inoculation of both young women and men with the HPV vaccine would markedly decrease HPV-associated oral cancers. - Remember the mouth heals quickly. If you have a sore in your mouth that doesn't heal in 14 days, seek prompt treatment from your dentist. - See your dentist at least once a year (better if twice a year) for a complete oral health screening. This is extremely important as oral cancer initially develops a precancerous oral lesions (patient treatment and survival is much better at this stage). Other social bookmarking and sharing tools: The above story is reprinted from materials provided by Ohio State University Center for Clinical and Translational Science, via Newswise. Note: If no author is given, the source is cited instead.
1.779754
HuggingFaceFW/fineweb-edu
be conducted safely) - also think about engaging parents and children in education resources such as e-bug and PHE schools resources - ensure parents and young people are aware of recommendations on transport to and from education or childcare setting (including avoiding peak times). Read the Coronavirus (COVID-19): safer travel guidance for passengers - talk to staff about the plans (for example, safety measures, timetable changes and staggered arrival and departure times), including discussing whether training would be helpful - communicate early with contractors and suppliers that will need to prepare to support your plans for opening for example, cleaning, catering, food supplies, hygiene suppliers - discuss with cleaning contractors or staff the additional cleaning requirements and agree additional hours to allow for this Keep cohorts together where possible and: - ensure that children and young people are in the same small groups at all times each day, and different groups are not mixed during the day, or on subsequent days - ensure that the same teacher(s) and other staff are assigned to each group and, as far as possible, these stay the same during the day and on subsequent days, recognising for secondary and college settings there will be some subject specialist rotation of staff - ensure that wherever possible children and young people use the same classroom or area of a setting throughout the day, with a thorough cleaning of the rooms at the end of the day. In schools and colleges, you may want to consider seating students at the same desk each day if they attend on consecutive days For cleaning and hygiene: - follow the COVID-19: cleaning of non-healthcare settings guidance - ensure that sufficient handwashing facilities are available. Where a sink is not nearby, provide hand sanitiser in classrooms and other learning environments - clean surfaces that children and young people are touching, such as toys, books, desks, chairs, doors, sinks, toilets, light switches, bannisters, more regularly than normal - ensure that all adults and children: - frequently wash their hands with soap and water for 20 seconds and dry thoroughly. Review the guidance on hand cleaning - clean their hands on arrival at the setting, before and after eating, and after sneezing or coughing - are encouraged not to touch their mouth, eyes and nose - use a tissue or elbow to cough or sneeze and use bins for tissue waste ('catch it, bin it, kill it') - ensure that help is available for children and young people who have trouble cleaning their hands independently - consider how to encourage young children to learn and practise these habits through games, songs and repetition - ensure that bins for tissues are emptied throughout the day - where possible, all spaces should be well ventilated using natural ventilation (opening windows) or ventilation units - prop doors open only if they are not fire doors, and where it is safe to do so (bearing in mind fire safety and safeguarding), to limit use of door handles and aid ventilation - get in touch with public sector buying organisation partners (for example ESPO, YPO, NEPO) about proportionate supplies of soap, anti-bacterial gel and cleaning products if needed - there is no need for anything other than normal personal hygiene and washing of clothes following a day in an educational or childcare setting Reduce mixing within education or childcare setting by: - accessing rooms directly from outside where possible - considering one-way circulation, or place a divider down the middle of the corridor to keep groups apart as they move through the setting where spaces are accessed by corridors - staggering breaks to ensure that any corridors or circulation routes used have a limited number of pupils using them at any time - staggering lunch breaks - children and young people should clean their hands beforehand and enter in the groups they are already in, groups should be kept apart as much as possible and tables should be cleaned between each group. If such measures are not possible, children should be brought their lunch in their classrooms - ensuring that toilets do not become crowded by limiting the number of children or young people who use the toilet facilities at one time - noting that some children and young people will need additional support to follow these measures (for example, routes round school marked in braille or with other meaningful symbols, and social stories to support them in understanding how to follow rules) Use outside space: - for exercise and breaks - for outdoor education, where possible, as this can limit transmission and more easily allow for distance between children and staff - although outdoor equipment should not be used unless the setting is able to ensure that it is appropriately cleaned between groups of children and young people using it, and that multiple groups do not use it simultaneously. Read COVID-19: cleaning of non-healthcare settings For shared rooms: - use halls, dining areas and internal and external sports facilities for lunch and exercise at half capacity. If class groups take staggered breaks between lessons, these areas can be shared as long as different groups do not mix (and especially do not play sports or games together) and adequate cleaning between groups between groups is in place, following the COVID-19: cleaning of non-healthcare settings guidance - stagger the use of staff rooms and offices to limit occupancy Reduce the use of shared resources: - by limiting the amount of shared resources that are taken home and limit exchange of take-home resources between children, young people and staff - by seeking to prevent the sharing of stationery and other equipment where possible. Shared materials and surfaces should be cleaned and disinfected more frequently - although practical lessons can go ahead if equipment can be cleaned thoroughly and the classroom or other learning environment is occupied by the same children or young people in one day, or properly cleaned between cohorts Adjust transport arrangements where necessary including: - encouraging parents and children and young people to walk or cycle to their education setting where possible - making sure schools, parents and young people follow the Coronavirus (COVID-19): safer travel guidance for passengers when planning their travel - ensuring that transport arrangements cater for any changes to start and finish times - making sure transport providers do not work if they or a member of their household are displaying any symptoms of coronavirus (COVID-19) - making sure transport providers, as far as possible, follow hygiene rules and try to keep distance from their passengers - taking appropriate actions to reduce risk if hygiene rules and social distancing is not possible, for example when transporting children and young people with complex needs who need support to access the vehicle or fasten seatbelts - local authorities or transport providers could consider the following: - guidance or training for school transport colleagues - substituting smaller vehicles with larger ones, or running 2 vehicles rather than one, where possible, to reduce the number of passengers per vehicle and increase the amount of space between passengers - cordoning off seats and eliminating face-to-face seating, where vehicle capacity allows, to help passengers spread out - communicating revised travel plans clearly to contractors, local authorities and parents where appropriate (for instance, to agree pick-up and drop-off times) Will contact tracing be in place in educational and childcare settings? The government has launched a new national test and trace programme. This will bring together an app, expanded web and phone-based contact tracing, and swab testing for those with potential coronavirus (COVID-19) symptoms. This programme will play an important role in helping to minimise the spread of coronavirus (COVID-19) in the future. It will also include more traditional methods of contact tracing if a child, young person or parent tests positive. This could include, for example, direct discussion with parents and schools or colleges on recent contacts. The government is recruiting 18,000 contact tracers to support contact tracing and will recruit more if needed. They will play an important part in tracing the contacts of those with coronavirus (COVID-19), including children. What happens if someone becomes unwell at an educational or childcare setting? If anyone in an education or childcare setting becomes unwell with a new, continuous cough or a high temperature, or has a loss of, or change in, their normal sense of taste of smell (anosmia), they must be sent home and advised to follow the COVID-19: guidance for households with possible coronavirus (COVID-19) infection guidance. If a child is awaiting collection, they should be moved, if possible, to a room where they can be isolated behind a closed door, depending on the age of the child and with appropriate adult supervision if required. Ideally, a window should be opened for ventilation. If it is not possible to isolate them, move them to an area which is at least 2 metres away from other people. If they need to go to the bathroom while waiting to be collected, they should use a separate bathroom if possible. The bathroom should be cleaned and disinfected using standard cleaning products before being used by anyone else. PPE should be worn by staff caring for the child while they await collection if a distance of 2 metres cannot be maintained (such as for a very young child or a child with complex needs). In an emergency, call 999 if they are seriously ill or injured or their life is at risk. Do not visit the GP, pharmacy, urgent care centre or a hospital. If a member of staff has helped someone with symptoms, they do not need to go home unless they develop symptoms themselves (and in which case, a test is available) or the child subsequently tests positive (see 'What happens if there is a confirmed case of coronavirus (COVID-19) in a setting?' below). They should wash their hands thoroughly for 20 seconds after any contact with someone who is unwell. Cleaning the affected area with normal household disinfectant after someone with symptoms has left will reduce the risk of passing the infection on to other people. See the COVID-19: cleaning of non-healthcare settings guidance. What
1.872349
openbmb/Ultra-FineWeb
The COVID19 pandemic has significantly disrupted the daily operations and strategic goals of civil society organisations and the non-profit sector. At the same time, many funders and donors have responded by shifting priorities towards emergency funds and short-term needs. As vaccines are being rolled out across Europe, it is time to reflect and prepare for what is coming next. The return to what has been before is unlikely. So how can organisations prepare and strategise for the transition to the post-COVID19 world? The interactive online discussion panel aims to open a dialogue between donor organisations and activists, civil society and change-making organisations on working together to meet future challenges effectively. The expert presentations aim to build capacity within civil society to help them prepare and manage the transition successfully. Organised with the support of the European Youth Forum. Programme and registration. Acompanhe o nosso Trabalho. subscrever newsletter
1.088068
Zyphra/Zyda-2
Nonlinear static and cyclic analysis of concrete-filled steel columns - Elsevier Ltd. - Publication Type: - Journal Article - Journal Of Constructional Steel Research, 2009, 66 (6), pp. 793 - 802 - Issue Date: Copyright Clearance Process - Recently Added - In Progress - Closed Access This item is closed access and not available. This paper presents the formulation of a one-dimensional (1D) composite frame element for the nonlinear static and cyclic analysis of concrete-filled steel (CFS) beamâcolumns. A two-node frame element is formulated using the force interpolation concept, and the material nonlinearity at section level is taken into account by employing a total secant stiffness and modified fibre element approach. The size effect and steel tube confinement on the concrete strength and ductility are taken into account. Further, the effect of steel tube local buckling on the member strength is addressed. Concerning geometrical nonlinearities at the element level, the equilibrium equations are satisfied for the deformed element to take account of PâÎ effects. The formulation accuracy and efficiency of the model are verified by some numerical examples of the static and cyclic analysis of CFS members. Please use this identifier to cite or link to this item:
1.086321
openbmb/Ultra-FineWeb
"Successful people ask better questions, and as a result, they get better answers." Tony Robbins. Investing time and energy into getting to know yourself more and developing yourself personally, spiritually and professionally can have massive returns. If you desire to achieve success of any kind, discovering and removing your personal barriers makes this journey worth it. Your state of mind and the emotions you feel have the power to shape how you experience life. Your inner world (mindset and emotions) makes all the difference between success and failure, satisfaction and losing oneself, strengthening and powerlessness, joy and profound stress/nervousness etc. History applauds individuals who have overcome difficult and challenging times, with courage, perseverance, and vigour. I can only imagine that these individuals whether consciously or unconsciously used questions to inspire their actions. Here are 12 questions and the benefits of answering them. Name something you are happy or are unhappy about and what decisions/choices you made that led to this result? - Take responsibility for your choices and what you manifest in your life. - Gain a clear understanding of the results you have created in your life - Examine what choices you have made that have led to the results What beliefs are restricting you from making progress? E.g., I am not good enough; I don't have time. - Investigate your underlying values and beliefs that either limit or assist you. - Discover what is important to you - Identify the beliefs that are assisting you or restricting you - Find out if you have shared values and habits with the people around you. Write down a general belief that you have about other people or the world, E.g., People can't be trusted, or life is what you make it. How does this affect you? - Become aware of what you project on the world and how this affects your perception of others and situations - Discover how you view and interpret what happens in your life - Think about the possibility of opposing views and other ways of seeing things. - Learn from what you think about others as something that exists in you. - Discover how your projections shape your daily experience; influencing your thoughts, feelings and actions. Name one thing that you would like to achieve in the next 12 months? - Develop goals and plan to move towards what you desire - Choose goals that will help you make progress in your life - Set goals that motivate you to take action. - Ensure your goals are Specific, Measurable, Achievable, Realistic, Time-bound – SMART Write down one habit that you would like to change and explain when this habit kicks in (the trigger is often unconscious), E.g., Getting angry when you feel powerless, constantly checking Facebook because you feel like you are missing out (FOMO.) - Gain knowledge of your triggers and what causes you to feel, think and act in a certain way. - Understand what causes you to think or feel a certain way - Become aware of what things or situations set your triggers off - Pay attention to when your triggers are set off - Learn how to observe the behaviour you display once your triggers have been set off – are these helpful or not? How do you feel on most days and what thoughts make you feel this way? e.g., Emotion = stressed; Thought = I can't cope - Take charge of your thoughts, feelings, words and actions - Become an observer of yourself - Learn which ones are constructive or disruptive - Discover ways to use them as indicators that you are on the right or wrong track - Create ways to positively program yourself into feeling good, e.g., affirmations What impact would you like to have in the world and why? What will your legacy be? Choose or discover your purpose, which will help to guide your career choices or how you can be of service. - Choose a purpose rather than trying to search for one - Create a vision based on your purpose outlining what you would like to achieve - Search for or create opportunities to help you experience your purpose - Identify how you can be of service to others. What makes you feel happiness and a sense of inner peace. How can you experience this more? - Experience more happiness, inner peace and joy in life - Discover what makes you feel happy, a sense of peace and joy - Discover what you allow to take these things away - Seek out or create ways to experience more of these - Learn to avoid or get rid of things that take them away. Which area of your well-being requires your attention? Mental, emotional, physical, spiritual. Name one thing you can do to improve. - Take care of your mental, emotional, physical and spiritual well-being - Challenge yourself to discover what prevents your well-being in each area. - Develop an understanding of what well-being looks and feels like. - Find practical ways to work on your well-being, e.g., fitness plan, healthy eating, meditation, etc. What motivates you to get things done, e.g., setting goals, writing a to-do list, telling other people - Improving efficiency and productivity - Discover why you put things off - Devise ways to increase your motivation and action - Learn how to prioritize what is essential and what you can delegate - Work harder and smarter to maximize your time Imagine that you were born to experience one thing, e.g., success, happiness, helping others, abundance, etc. Write down the experiences you would like to have. - Shape your life experience - Discover what you would like to experience in life, e.g., success, wealth, making a difference, etc. - Identify ways to create the experiences you desire. - Learn how to overcome any obstacles that prevent you from these experiences. What are you confident at and say why? What are you not confident at and say why? - Increase your confidence - Learn what prevents you from feeling confident - Discover the beliefs that hold you back - Learn to believe in your ideas, abilities - Become brave and live your life as you choose, as long as it doesn't hurt you or others. By Dionne Jude - Co-Founder Overstand You
1.684757
openbmb/Ultra-FineWeb
"Physicians can greatly help to increase back sleeping by advising parents and caregivers to place infants to sleep on their backs," said Dr. Duane Alexander, M.D., Director of the National Institute of Child Health and Human Development (NICHD), which funded the study. The analysis also found that the number of infants being put to sleep on their stomachs continues to drop. But a significant portion of the population is still placing infants to sleep on their stomachs-despite having been told that back sleeping reduces the risk of SIDS. Marian Willinger, Ph.D., of NICHD's Pregnancy and Perinatology Branch, led the research team that reported the findings, which appear in the April 26 Journal of the American Medical Association. Other members of the research team are from the National Institute on Deafness and Other Communication Disorders (NIDCD), the Harvard School of Medicine, and the Boston University School of Medicine. SIDS strikes nearly 3,000 babies each year. SIDS is the sudden, unexplained death of an infant in the first year of life. The causes of SIDS are still unclear, and it is not currently possible to predict which infants might fall victim to SIDS. Back sleeping has been found to significantly reduce the risk of SIDS. Since the dissemination of the recommendation to place infants to sleep on their backs or sides, the SIDS rate has fallen almost 40 percent. The NISP study is an annual telephone survey of nighttime caregivers of infants under 8 months of age conducted in the 48 contiguous states to assess infant care practices and to help evaluate the dissemination of the Back to Sleep Campaign messages. This campaign, which encourages that infants be placed to sleep on their backs to reduce the risk of SIDS, was launched in 1994 by the NICHD, in partnership with other agencies of the U.S. Public Health Service, the AAP, the SIDS Alliance, and the Association of SIDS and Infant Mortality Programs. This report examines the factors that determine sleep position choice between 1994 and 1998. The study found that physicians had the greatest influence on whether or not caregivers placed infants to sleep on their backs. In fact, caregivers that were advised by physicians to put infants on their backs to sleep were about three times more likely to do so than were caregivers not given a recommendation. "Unfortunately, 41 percent of respondents said that their physicians had not made a recommendation on sleep position in _PHONE_," the researchers wrote. A variety of other sources also helped to increase the rate of back sleeping. The study found that in _PHONE_, about 79 percent of caregivers said they had received the recommendation to put infants to sleep on their backs or sides from at least one of four sources: physician, nurse, reading materials, or radio and television, compared with 38 percent in 1994. Also, by _PHONE_ only 5.8 percent of caregivers said they had been told by at least one source to place infants to sleep on their stomachs, compared with 24.4 percent in 1994. Caregivers who reported that they received a recommendation of back sleep position from all four sources were six times more likely to place the infant on the back. Significant racial disparities remain in sleep position choice. Between 1994 and 1998, stomach placement had decreased from 44 percent to 17 percent among white infants, and from 53 to 32 percent among black infants. African American infants are 2.2 times more likely to die from SIDS than are white infants. Stomach sleeping was not confined to infants whose caregivers had not heard the Back to Sleep message. Indeed, from 1995 to 1998, 86 percent of those caregivers who placed their infants on the stomach reported receiving only recommendations to place them on their backs or sides. "In contrast, within a subgroup of caregivers, there appears to be a strong motivation to place the infant prone that overrides the health education message," the authors wrote. Specifically, caregivers who put infants to sleep on their stomachs said that the baby liked it or slept better that way. They also feared the infants might choke on their own vomit. The researchers added that studies in England and Australia had shown that there were no increases in breathing problems that might be due to choking for infants placed on their backs. In contrast, both studies had found that parents reported that there were more breathing problems in infants placed to sleep on their stomachs. "Motivation strongly influences the choice of prone position in the current environment; thus clinicians should discuss the reasons for choice of sleep position with the caregiver," the authors To help eliminate the racial disparity in the SIDS rate between African American and white infants, the NICHD is developing a community-based outreach program in collaboration with African American organizations to extend the Back to Sleep message into diverse communities. NICHD and the NIDCD are both institutes of the National Institutes of Health, the Federal government's premier biomedical research agency. NICHD supports and conducts research on the reproductive, neurobiological, developmental, and behavioral processes that determine and maintain the health of children, adults, families, and populations. The NICHD website, _URL_ contains additional information about the Back to Sleep Campaign, the institute, and its mission.
2.01602
HuggingFaceFW/fineweb-edu
secondary probe temperature plus the permissible DUT stress, then the CSP is reduced such that it is equal to the secondary probe temperature plus the permissible DUT stress at that temperature. This "final" CSP value is then passed to the temperature control system. - [0076] Detailed Description of Method and Associated Algorithm - [0077] A detailed description of the method 300 of FIG. 3 is now provided with reference to FIGS. 3 a through 3 d, and the definitions and assumptions provided herein. While the following discussion is cast in terms of the method employed within the Applicant's "Intelligent 2 Probe Control" (hereinafter "I2PC") computer program embodiment, it will be recognized that other algorithms, firmware, or even hardware embodiments of the disclosed method may be used with equal success. It is also noted that while the terms "determine" and "calculate" are used in describing the following method, these terms are not meant to be limited to specific processes. For example, it is contemplated that in lieu of calculating a specific value, such value may be provided by the DUT or TCS manufacturer, or otherwise obtained without the need for explicit calculation. - [0078] In the first process step 302 (FIG. 3 a), an allowable or System Operating Range (SOR) is determined for the temperature control system. This process step 302 is comprised of several sub-steps 304, 306, 308, as follows. In sub-step 304, a system lower aggregate operating limit (LAOL) is determined as being the higher of the following: (a) the low limit of the system temperature controller; (b) the low limit of the temperature equipment; or (c) the low limit of the DUT. Similarly, in sub-step 306, a system upper aggregate operating limit (UAOL) is determined to be the lower of: (a) the high limit of the system temperature controller; (b) the high limit of the temperature equipment; or (c) the high limit of the DUT. Finally, in sub-step 308, the system operating range (SOR) is defined as the range between and including the LAOL and the UAOL determined in sub-steps 304 and 306. - [0079] Note that in the present embodiment, a valid SOR is defined as one where the LAOL is a lower temperature than the UAOL. If this condition is not met, the algorithm exits with and generates an appropriate error code. - [0080] Next, in the second process step 310 (FIG. 3 b), the allowable or DUT operating range (DOR) is determined by calculating the DUT permissible stress at the current secondary probe temperature. In the first sub-step 312 of the second process step 310, the DUT low limit permissible stress is subtracted from the DUT high limit permissible stress to determine the DUT stress range. In sub-step 314, the DUT low limit temperature is subtracted from the DUT high limit temperature to determine the DUT range in degrees. Next, the secondary probe temperature is subtracted from the DUT high limit temperature in sub-step 316. The percentage of the DUT temperature range represented by the secondary probe temperature is then calculated in sub-step 318 by dividing the result of sub-step 316 by the result of sub-step 314. The DUT currently permissible stress (DCPS) is determined at the current secondary probe temperature (e.g., that of the DUT core) by multiplying the result of sub-step 312 by the result of sub-step 318 and subtracting this product from the high limit permissible stress for the DUT in sub-step 319. - [0081] Next, the DUT upper operating limit (DUOL) is calculated by adding the DCPS to the current secondary probe temperature (DUT core) in sub-step 320. Similarly, the DUT lower operating limit (DLOL) is determined by subtracting the DCPS from the current secondary probe temperature in sub-step 322. Finally, in sub-step 324, the DOR is defined as the temperature range between arid including the DUOL and the DLOL. - [0082] In the third process step 326 (FIG. 3 c), the system parameters are evaluated to determine if the I2PC algorithm can operate to generate a valid CSP. It should be noted that in the present embodiment, some portion of the DOR must overlap a portion of the SOR in order for the limits of all devices to be respected. If the core temperature of the DUT plus or minus the permissible stress at that temperature defines a range (e.g., the DOR) that is outside the range that is defined by the SOR (e.g., the limits of the DUT, equipment, and temperature control system) then it will not be possible to determine a setpoint that is within both ranges and thus which respects the limits of both the DUT permissible stress and those associated with the remainder of the temperature control system. - [0083] Referring again to FIG. 3 c, the DUOL and DLOL are each compared to the UAOL and LAOL in sub-steps 328 a, 328 b and sub-steps 329 a, 329 b, respectively, of process step 326. If either: (I) the DUOL is less than the UAOL and greater than the LAOL; or (ii) the DLOL is greater than the LAOL and less than the UAOL, then a valid CSP may be calculated by the algorithm. If neither of these conditions are met, then an error code is generated by the algorithm. It will be appreciated that while a parallel approach to these comparisons is illustrated in FIG. 3 c, other methods of comparison and logical relationships may be substituted. - [0084] In the fourth process step 330 of the method 300 (FIG. 3 d), the Control Setpoint (CSP) is calculated. Initially, the secondary probe value is subtracted from the PSP, and the absolute value of this quantity taken in sub-step 332. In sub-step 334, the aforementioned absolute value is multiplied by a first setup factor (F34). See Appendix A. In step 336, the value of a second setup factor (F35) is added or subtracted as appropriate to the result. The first setup factor acts as a scaling factor or multiplier for the proportional term of the CSP, while the second factor represents a thermal overdrive value (in degrees). In the present embodiment, the first and second setup parameters are numerical values pre-selected or input by the operator, although it can be appreciated that these parameters can be supplied dynamically during the temperature conditioning process from another algorithm or source if desired. Typical values for the first setup factor F34 are in the range of 0.0 to 5.0 (default value=2). Typical values for the second setup factor are −20.0 to +20.0° C. (default=5.0° C.), or −36.0 to +36.0° F. (default 9° F. The foregoing values are merely illustrative; other values may be chosen. - [0085] In sub-steps 338 and 339, the result of sub-step 336 is compared to zero and if greater than zero, is added to the PSP to determine the so-called "unlimited" CSP. Next, the result of sub-step 339 is compared to the UAOL and the DUOL in sub-step 340. The CSP is set to the lesser of these three values (e.g., unlimited CSP, UAOL, and DUOL). Lastly, the result of sub-step 340 above is compared to the LAOL and the DLOL per sub-step 342. The CSP is then set to the greater of these three values. This is the "final" CSP. - [0086] In the final process step 344 of the method 300 of FIG. 3, the "final" CSP is passed to the temperature control system for use thereby. - [0087] It will further be recognized that the method 300 (and associated algorithm) disclosed herein has several operational attributes which provide advantages over prior art systems and methods. Specifically, the algorithm of the present invention (i) automatically reduces the excess heating/cooling as the DUT core approaches the PSP; and (ii) automatically stops moving the setpoint and enters into normal PID control when either the PSP is reached (within the tolerance of the settling band temperature tolerance parameter F31, described below), or when a predetermined period of time without significant change in the DUT core temperature expires. These attributes are discussed in additional detail below. - [0088] Automatic Reduction of Excess Heating/Cooling—Because the CSP exceeds the PSP by an amount related to the difference between the secondary probe and the PSP, the amount the CSP leads the PSP is automatically reduced as the DUT core temperature approaches the PSP. Using the CSP to cause the thermal environment to exceed the PSP results in faster thermal transfer to/from the DUT to the increased differential. Assuming that the DUT skin temperature approaches the environment temperature, the DUT core is the primary beneficiary of the increased heat transfer. - [0089] As the DUT core continues to increase in temperature due to this increased differential, the secondary probe temperature
1.69145
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
Definitions for uterusˈyu tər əs; ˈyu təˌraɪ This page provides all possible meanings and translations of the word uterus Princeton's WordNet 1. uterus, womb(noun) a hollow muscular organ in the pelvic cavity of females; contains the developing fetus Wiktionary 1. uterus(Noun) An organ of the female reproductive system in which the young are conceived and develop until birth; the womb. 2. Origin: From uterus. Webster Dictionary 1. Uterus(noun) the organ of a female mammal in which the young are developed previous to birth; the womb 2. Uterus(noun) a receptacle, or pouch, connected with the oviducts of many invertebrates in which the eggs are retained until they hatch or until the embryos develop more or less. See Illust. of Hermaphrodite in Append Freebase 1. Uterus The uterus or womb is a major female hormone-responsive reproductive sex organ of most mammals including humans. One end, the cervix, opens into the vagina, while the other is connected to one or both fallopian tubes, depending on the species. It is within the uterus that the fetus develops during gestation, usually developing completely in placental mammals such as humans and partially in marsupials such as kangaroos and opossums. Two uteri usually form initially in a female fetus, and in placental mammals they may partially or completely fuse into a single uterus depending on the species. In many species with two uteri, only one is functional. Humans and other higher primates such as chimpanzees, along with horses, usually have a single completely fused uterus, although in some individuals the uteri may not have completely fused. In English, the term uterus is used consistently within the medical and related professions, while the Germanic-derived term womb is more common in everyday usage. Most animals that lay eggs, such as birds and reptiles, including most ovoviviparous species, have an oviduct instead of a uterus. Note however, that recent research into the biology of the viviparous skink Trachylepis ivensi has revealed development of a very close analogue to eutherian mammalian placental development. U.S. National Library of Medicine 1. Uterus The hollow thick-walled muscular organ in the female PELVIS. It consists of the fundus (the body) which is the site of EMBRYO IMPLANTATION and FETAL DEVELOPMENT. Beyond the isthmus at the perineal end of fundus, is CERVIX UTERI (the neck) opening into VAGINA. Beyond the isthmi at the upper abdominal end of fundus, are the FALLOPIAN TUBES. Translations for uterus From our Multilingual Translation Dictionary Get even more translations for uterus » Translation Find a translation for the uterus definition in other languages: Select another language: Discuss these uterus definitions with the community: Word of the Day Would you like us to send you a FREE new word definition delivered to your inbox daily? Please enter your email address: Citation Use the citation below to add this definition to your bibliography: Style:MLAChicagoAPA "uterus." Definitions.net. STANDS4 LLC, 2015. Web. 3 Sep. 2015. <_URL_ Are we missing a good definition for uterus? Don't keep it to yourself... Nearby & related entries: Alternative searches for uterus: Thanks for your vote! We truly appreciate your support.
2.569954
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
How is SOHO Orbit Location (The L1 point) Calculated? (by Amara Graps) SOHO has an uninterrupted view of the Sun. That view of the Sun is achieved by operating SOHO from a permanent vantage point 1.5 million kilometers (about 4 times as far away as the moon) ahead of the Earth in a "halo orbit" around the L1 Lagrangian point, where it has a pointing accuracy of 10". All previous solar observatories have orbited the Earth, where their observations were periodically interrupted as our planet 'eclipsed' the Sun. Look at an orbital schematic for SOHO's halo orbit. The halo orbit has a period of 180 days. The semi-diameters of the halo orbit have typical dimensions of approximately 200,000 km in the ecliptic in the Earth-Sun direction, 650,000 km in the ecliptic perpendicular to the Earth-Sun direction, and 200,000 km out of the ecliptic. The L1 halo orbit has several particular advantages as compared to low Earth orbits: - It provides a smooth Sun-spacecraft velocity change throughout the orbit, appropriate for helioseismology. - It is permanently outside of the magnetosphere, appropriate for the ''in situ'' sampling of the solar wind and particles. - It allows permanent observation of the Sun, appropriate for all the investigations. You may jump to Goddard's Exploration WWW Site for a nice intuitive, elementary description about the Lagrangian points. If you wish to read a more advanced description of the Lagrangian points, then please read on. The calculation of Lagrange points can be found in books on orbital or celestial mechanics. One of my favorites is:Fundamentals of Celestial Mechanics by J.M.A. Danby, Willmann-Bell, 1988.In particular, look at your celestial mechanics textbook under the category of the Three-Body Problem, usually something like "Positions of Equilibrium" or "The Libration Points." (In the above text, this subject is in Chapter 8). Also check out: - Orbit Visualization Project Home page, in particular: The Three Body Problem - Computing the Libration Points. (They considered the Circular Restricted Three-Body Problem or "CR3BP") (Information from the Danby book and the site mentioned above.) The general problem of the motion of three bodies (assumed to be point masses) subject only to their mutual gravitational attractions has not been solved analytically. Think about this. It's only slightly more complicated than the two-body problem! Nevertheless, if one imposes some special constraints on the three-body problem (the so-called "restricted three-body problem"), some particular solutions can be derived, and it has important applications to actual celestial objects (consider the Sun-Earth-Moon system or a Saturn-moonlet-ring particle system). The restricted three-body problem is where two bodies of a finite mass revolve around one another and a third body of infinitesimal mass moves in their field. The third object essentially is unable to measurably affect the motion of the two other bodies. Since the orbits of the two finite mass bodies conform to the known solution of the two-body case, the problem is reduced to the study of the motion of only the small test particle in the field of the two coorbiting primaries. In order to simplify the problem even further, we can consider the *circular* restricted three-body problem, in which the two finite masses are assumed to move in circular orbits about one another. There is also the *elliptic* restricted three-body problem, which is the topic of considerable research activity. If one sets up the circular restricted three body problem so that the center of mass of the two finite objects is the origin and the axes rotate with constant angular velocity, the first thing to fall out of one of the constants of integration is an energy integral called "Jacobi's integral" (after the German mathematician Carl Gustav Jacob Jacobi), and it represents the conservation of energy in the rotating system. Five additional integrals are required for the restricted three body problem's complete solution, but none has ever been found, even though people have cleverly tried to invoke things like conservation of angular momentum to get a second integral. So assume that we now have the Jacoby Integral. We want to know the positions of equilibrium of this system. There exist five fixed points in the CR3BP (Circular Restricted Three-Body Problem). Also known as Lagrangian, or Libration points, these points represent points in space where an unmoving particle would remain for all time. These fixed points are all points where all six phase variables time derivatives are zero. In other words, the libration points, are solutions to the equations of motion for the spacecraft such that the x, y, and z positions and x, y, and z velocities are all 0. To compute the location of these points, one must solve the equations of motion under these special conditions. When solving, one finds that there are 5 solutions to these equations: - Two solutions fall out easily; each forms and equilateral triangle with the Earth and the Sun in the plane of rotation. - Each of the 3 remaining solutions also lie in the plane of rotation, on the line between the Earth and Sun. Two of these points, known as L4 and L5 form equilateral triangles with the primary masses. The other three libration points (L1, L2, L3) are colinear with the primary masses, and are found by solving quintic polynomials dictated by the equations of motion. These solutions are diagrammed at the Geometry Web Site. To find the location of the 3 libration points lying on the Earth-Sun axis (L1, L2, and L3), one must solve for the x coordinate of each. This boils down to solving a quintic polynomial equation. The polynomial equation is the following. Let V be a modified potential of this CR3BP system: -V = (1/2)(x^2 + y^2) + (1-mu)/r1 + mu/r2 (1)where mu is the ratio of the smaller mass to that of the larger mass (3.03591e-6 for the Earth-Sun system) r1 and r2 are the distances from the massless particle to mass1 and mass2. The surface of zero relative velocity is x^2 + y^2 + 2*(1-mu)/r1 + 2*mu/r2 = Constant (2)The surface of zero relative velocity is given by the function: F(x,y,z) = V + 1/2*Constant = 0. (3) There are two locations of equilibrium that a particle will move to, if given a nudge. You can see graphics of the two stable points here. So let's say a particle is given a nudge. It will go to one of two locations. The condition for the two places is: dF/dx = dF/dy = dF/dz = 0 (4)(all "d"s are partial derivatives) and we have from our relationship for our modified potential (1) and the function F: (3) dF/dz = z((1-mu)/r1^3 + mu/r2^3) (5)Any double point in the xy plane is where z = 0, so using the conditions for dF/dx=0 and dF/dy=0 (equations 4) we get x - (1-mu)*(x-x1)/r1^3 - mu*(x-x2)/r2^3 = 0 (6) y - (1-mu)*y/r1^3 - mu*y/r2^3 = 0 (7) Equation (7) is satisfied if y = 0 (Remember we are seeking a quintic polynomial along the x-line) and so equation (6) becomes: x - (1-mu) * (x-x1)/ |x-x1|^3 - mu*(x-x2)/ |x-x2|^3 = 0 (8)***where | | indicates a positive value of the difference. There are 3 real roots of the above equation along the x-axis. These correspond to the Lagrangian points L1, L2, and L3. Remember that around L1 is where SOHO is located in a Halo orbit. (If you would like to continue this exercise to find the Lagrangian L4 and L5 positions, then set y not equal to 0 in equation (7).) The Geometry Web page expands Equation (8) out for L1 into: -((1-mu)(1-mu+x)^2) + mu(mu+x)^2 + x(1-mu+x)^2(mu+x)^2 = 0 And you can see curves for the equations for L1 corresponding quintic polynomials for several values of mu. Now what about the Halo Orbits? Once the Libration points have been located, one can search for orbits about these points. Surrounding each of the three collinear libration points (L1, L2, L3) exist families of periodic orbits. These periodic orbits, known as "halo orbits", provide excellent outposts for astronomical observation. So if we look out at one of the libration points, say the L1 point, we would find the spacecraft orbiting it in a small circle that resembles a halo around L1. Halo orbits' proximity to the Earth, and accesiblity to low-velocity space craft provide an economical opportunity to expand out knowledge about our solar system. Furthermore, the existence of stable
2.270915
HuggingFaceFW/fineweb-edu
This includes access to industry reports, market research, and other resources that can help users to stay ahead of the competition and identify new opportunities for growth and innovation.Overall, GRANAT88 is a powerful platform that provides individuals and businesses with the tools and resources they need to embrace the future and stay ahead of the competition. Whether you are an entrepreneur looking to launch a new product or service, or a business leader looking to stay ahead of the curve, GRANAT88 can help you to achieve your goals and unlock your full potential.In conclusion, innovation is the key to success in today's fast-paced business environment, and GRANAT88 is the ultimate guide to innovation. By embracing the platform's collaborative, creative, and forward-thinking approach, individuals and businesses can unlock their full potential and achieve unprecedented levels of success and growth. Togel, a popular form of gambling, has been played in various parts of the world for centuries. It originated in China and gradually spread to other parts of Asia, including Indonesia, where it is now commonly referred to as "IDN Togel". In recent years, the emergence of online platforms has made it possible for people to play IDN Togel from anywhere in the world. However, for newcomers to the game, the world of IDN Togel can be quite confusing. This article provides a comprehensive guide to help you decode the world of IDN Togel.IDN Togel is a type of lottery game that involves selecting a set of numbers from a pool of numbers. The objective is to correctly predict the winning numbers. The game is played with a variety of different formats, including 2D, 3D, 4D, and 6D. Each format requires the player to select a different number of digits. For example, in the 2D format, the player selects two digits, while in the 6D format, the player selects six digits. The more digits the player selects, the higher the potential payout.IDN Togel is played by millions of people worldwide, and it is particularly popular in Indonesia. There are many online platforms that offer IDN Togel games, and players can choose from a wide range of different GRANAT88 formats and betting options. One of the main advantages of playing IDN Togel online is that it is much more convenient than playing in a physical location.
1.181548
openbmb/Ultra-FineWeb
DETERMINATION OF REGULATORY RELIABILITY FOR DEFORMATION OF REINFORCED-CONCRETE FLEXIBLE STRUCTURES Vestnik Tomskogo gosudarstvennogo arkhitekturno-stroitel nogo universiteta JOURNAL of Construction and Architecture Strength, stiffness, crack resistance requirements must be specified for any structure. These requirements are provided by the appropriate calculations. Compliance with all requirements ensures the design reliability. But due to the fact that structural analysis is performed by the limit state method, the numerical value of the reliability cannot be determined. This can be done only by applying probabilistic methods of calculation. One of the major obstacles to the use of probabilistic methods ... babilistic methods is the lack of standardized values of the structural reliability (normative reliability).Although probabilistic methods of calculation were developed in late in the 20th century, the assignment of normative reliability values encountered serious obstacles. The most serious obstacle was the consideration of the failure consequences as some consequences were quantifiable. On the other hand, much work concerned the assignment of normative structural reliability based on their bearing capacity. Only few works related to the reliability normalization using deformation and crack resistance.This paper deals with the regulatory definition of reliability using the deformation of reinforced-concrete bendings based on physiological requirements. A method for determining the normative reliability based on human sensitivity to fluctuations is proposed. The normative reliability of precast slabs is determined using the proposed method.
1.595152
openbmb/Ultra-FineWeb
September 3 Birthday Horoscope september 3 birthday horoscope Can we concur on a few points like september 3 birthday horoscope? Putting your feet on the ground can attach you back to Mother Planet and assist center you. Again, a reality. Dark delicious chocolate transcends to milk delicious chocolate. Ehh … this could not be a reality depending on your tastebuds, however we would certainly love to assume that it holds true taking into consideration all the health advantages. Seriously, one truth that we all know is real is that we live in a world of duality, meaning that there are constantly two sides to whatever. This truth can be proven with anything. Believe about exactly how a plant expands: there's the gorgeous development that we see over ground, however additionally intricately linked roots that grow below the ground. Believe regarding how you yourself exist in this realm. You have two sides, as well. There are both rational and psychological aspects of your nature. You have actually left and right sides of your brain. And everyone has both manly and feminine energies. Getting to understand both sides of on your own from a sensible and psychological point of view can assist you to recognize just how to show up the life that you absolutely desire to live. Take Advantage Of Logic Using Ancient Scientific Devices Like Astrology Think it or otherwise, astrology is based a lot more on scientific research that's backed by facts as opposed to necromancy. september 3 birthday horoscope This is so much so that a selection of ancient cultures have used it's rational thinking to in fact compute schedules, determine substantial choices when it concerned battle strategy, and also planned their plants around astrological days. Astrology uses a system that integrates math and astronomy to create an intimate language that just a couple of recognize on a really deep level. Not only do astrologists formulate graphes (similar to mathematical charts), they also research particular aspects and transportations that integrate levels (these can look like geometric angles that you might have examined in math class). To come to be a fluent astrologer, you have to dedicate to both the usage of scientific research and math, but additionally use your instinct! Integrating Psychological Knowledge To Open Your Heart Course That's simply not real. Everybody can get in touch with their instinct and should! The feelings are indispensable to being familiar with on your own on a deeper, spiritual level. Actually, we would certainly presume as to state that feelings are exactly how your heart talks to you, but you need to train yourself to pay attention. The emotions that all of us have access to is actually a guidance system. It's important to actually experience each emotion so that you can be guided to where you're in fact intended to be. This is just how you take advantage of your intuition. Yes, it's excellent to really feel mad occasionally. Yes, it's good to really feel depressing often. These are all valid feelings that offer proof and advice as to where your soul would certainly like you to go and where your soul would not like you to go. When something makes you worried, that is your spirit talking to you, alerting you that maybe that isn't the course for you. While when something makes you ecstatic and energetic, your soul is saying loudly, "YES!" and encouraging you to progress. Recognizing the feelings is an essential step right into entering your spiritual power and brightening your heart course. How To Utilize Both Logic and Emotion to Materialize the Life Your Spirit Truly Desires september 3 birthday horoscope, as we discussed above, we live in a realm that is double in many means. Astrology is one the earliest, most trusted devices that humanity has had the ability to utilize to recognize the spirit's function by very first evaluating the Sun indication. The sun indication is your individual zodiac indicator that actually outlines your one-of-a-kind qualities and also highlights your emotional triggers and exclamations. Not just can you utilize a scientific tool that's packed with logic, but while exploring your personal astrology, you can also practice using your emotions. How does the info offered to you make you really feel? The response is usually reassured and reenergized to materialize the life that your heart is yearning for. Read your Individualized Soul Course Report right currently. The concern is, are you all set to touch right into both your logic and feelings? You've heard of NLP? It's nearly cultish. It represents neuro-linguistic programs and it resembles someone had actually taken the most effective out of all types of psychiatric therapy, tossed them in a mixer and developed NLP. It is a method of reasoning. It's a structure for exactly how to approach your life to be a lot more effective. It resembles a viewpoint of life based upon understanding just how your mind features. Second, it is a psycho-therapy device. It's utilized to deal with phobias and to change ideas in individuals. Some consider it as effective as CBT (cognitive-behavioral treatment) and most claim it's a whole lot faster than the Freudian college of thought. Brain training vector image. Depression stress symbols. Mental health and wellness concept.Because at the core of NLP exists hypno-therapy. To be much more particular, the "behavioral transforming" component of NLP is improved the teachings of Milton Erickson, one otherwise one of the most famous hypnotherapists of all times. And you recognize what hypno-therapy does? It's an accessibility, a hack, to your subconscious mind. Let me give you this example. Visualize your mind being like a computer system. Your conscious mind is accessing programs, turning it on and off and so forth. september 3 birthday horoscope, Regular customer access. Your subconscious mind is like going into the setups, setting up applications, removing them, including new customers and so forth. It's the "engine" or the "back room" that controls every little thing. Well, NLP is like a cyberpunk that can damage the password of your subconscious mind and modification there. It allows you to transform beliefs and to alter who you are, as a person, to your core. It goes to those deep thoughts and ideas you don't even understand you have and makes you act differently by tweaking your standard concepts of self. And with the usage of NLP you can remove fears, you can alter basic preferences (as not liking delicious chocolate anymore) or you can even heal wounds from the past, as those from childhood. It's efficient. It works well, and it's recommended by some of the best-known authors on the planet. Tony Robbins built his job and success on NLP and if you run a survey amongst successful individuals, you'll see that most of them are NLP practitioners or masters. You can't mess with what works. So why am I informing you this? If NLP resembles a hacker that tweaks your subconscious mind so you can become the individual you intend to be, if NLP is like a layout device that enables you to revamp that you are, after that brainwave entrainment is a device that makes it quicker and easier. And what does this mean for you? It's basic. Use them both. Make use of the impressive power of NLP to transform beliefs and self-identity ideas and use the power of brainwave entrainment to assist your subconscious mind be receptive and open. Open the door with brainwave syncing and change the furniture with NLP. A straightforward method to comprehend this is songs. If it's an unfortunate track, you end up being sad. If it's a lively track, it boosts your energy. If you listen it lots of times, you wind up assuming like in the track. Music is a wonderful instance of a tool that accesses your subconscious mind without also understanding. You switch on YouTube and through rep and emotion, you wind up being one with the music. Have you attempted reading a book you had no passion in analysis? Maybe for college or work? You read the words; they are verbalized in your mind but you finish up neglecting everything. They "enter one year and departure another". This is due to the fact that your subconscious mind is as shut as it obtains and learning can not take place in your aware dimension. This is why I use NLP in all my devices. Brainwave entrainment is remarkable for helping your mind run at that greater frequency. They're like Vitamin C for your life, making whatever better. And if you use brainwave entrainment, this is sufficient to attract, to show up advantages in your life since you'll be running at a higher degree. … september 3 birthday horoscope True, real change long-term not by attracting things bring in our lives but by changing who altering are so those things become a come to be orderAll-natural It's when we transform our ideas and concepts so we transform right into individuals for whom success and joy are 2 natural things. And this is where NLP functions so well. It helps whatever from healing fears of spiders to surpassing PTSD (post-traumatic anxiety disorder) to program you to be a lot more positive in front of a crowd or to be much more assertive with your colleagues. This is since while temporary victories are terrific– if you desire an effective life, you should come to be the person that is normally successful. And the only difference in between you and an effective individual is that the other one created the ideas and identification of such an individual. If you establish them as well, there's nothing quiting you from doing simply. I develop my program "Symptom Magic" around NLP. Each audio consists of concealed ingrained commands that reprogram your subconscious mind. Like a good hacker, brainwave syncing is opening the path
1.321924
Zyphra/Zyda-2
for National Geographic News Bacteria can survive in deep freeze for hundreds of thousands of years by staying just alive enough to keep their DNA in good repair, a new study says. In earlier work, researchers had found ancient bacteria in permafrost and in deep ice cores from Antarctica. These bacteria, despite being trapped for millennia, were able to be revived and grown in the lab. Some researchers had thought that bacteria would have to turn into dormant spores to survive for so long. But if bacteria merely went dormant, metabolism would stop and various environmental factors would begin damaging their DNA. Like an ancient scroll that's crumbling apart, the DNA becomes so damaged that it's indecipherable after about a hundred thousand years. Then the cells can't ever reproduce and the bacteria are effectively dead. "Our results show that the best way to survive for a long time is to keep up metabolic activity," said Eske Willerslev, lead study author and a researcher at the University of Copenhagen in Denmark. Doing this "allows for continuous DNA repair," Willerslev added. The work suggests that if bacterial life existed on Mars or on Jupiter's moon Europa, it might still survive locked in icy soils. The new study appears this week in the online advance edition of the Proceedings of the National Academy of Sciences. Living, Just Barely SOURCES AND RELATED WEB SITES
2.427335
HuggingFaceFW/fineweb-edu
following a single gene therapy administration, and its combination with sTGFβR2 reduced kidney atrophy by 75% in mice with renal fibrosis. Heart function in mice with heart failure improved by 58% when they were given sTGFβR2 alone or in combination with either of the other two genes, showing that a combined therapeutic treatment of FGF21 and sTGFβR2 could successfully treat all four age-related conditions, therefore improving health and survival. Administering all three genes together resulted in slightly worse outcomes, likely from an adverse interaction between FGF21 and αKlotho, which remains to be studied. A single combination gene therapy treats multiple age-related diseases Comorbidity is common as age increases, and currently prescribed treatments often ignore the interconnectedness of the involved age-related diseases. The presence of any one such disease usually increases the risk of having others, and new approaches will be more effective at increasing an individual's health span by taking this systems-level view into account. In this study, we developed gene therapies based on 3 longevity associated genes: fibroblast growth factor 21 (FGF21), αKlotho, soluble form of mouse transforming growth factor-β receptor 2 (sTGFβR2). The gene therapies were delivered using adeno-associated viruses, and we explored their ability to mitigate 4 age-related diseases: obesity, type II diabetes, heart failure, and renal failure. Individually and combinatorially, we applied these therapies to disease-specific mouse models and found that this set of diverse pathologies could be effectively treated and in some cases, even reversed with a single dose. We observed a 58% increase in heart function in ascending aortic constriction ensuing heart failure, a 38% reduction in α-smooth muscle actin (αSMA) expression, and a 75% reduction in renal medullary atrophy in mice subjected to unilateral ureteral obstruction and a complete reversal of obesity and diabetes phenotypes in mice fed a constant high-fat diet. Crucially, we discovered that a single formulation combining 2 separate therapies into 1 was able to treat all 4 diseases. These results emphasize the promise of gene therapy for treating diverse age-related ailments and demonstrate the potential of combination gene therapy that may improve health span and longevity by addressing multiple diseases at once. GDF11 as a Calorie Restriction Mimetic GDF11 was one of the first factors in blood identified as a possible explanation for the outcome of heterochronic parabiosis. When a young and old mouse have their circulatory systems joined, some aspects of aging reverse in the old mouse, and some aspects of aging are accelerated in the young mouse. GDF11 levels decline with age, and it was thought that increased levels of GDF11 provided by the young animal could act to improve function of cells and tissues in the older animal - though it was not well understood as to how GDF11 worked to produce these results. Since then there has been some debate over whether or not the original GDF11 research was technically correct, as well as some debate over whether or not factors in young blood are in fact responsible for parabiosis effects. Researchers have demonstrated benefits in old mice by delivery of GDF11 as a treatment, however. A company, Elevian, was founded to carry forward the development of GDF11 as basis for clinical therapy. Meanwhile, research into GDF11 and aging continues elsewhere in the scientific community. Today's very interesting open access research provides evidence for GDF11 to produce benefits in large part through triggering many of the same mechanisms as calorie restriction. If this is the primary mechanism of action, it would make GDF11 much less interesting as a basis for human therapy. Firstly because calorie restriction already exists, and is essentially free, and secondly because the practice of calorie restriction produces much larger effects on life span in short-lived species than it does in long-lived species. A blood factor involved in weight loss and aging In a previous study using mouse models, scientists observed that injecting aged mice with blood from young mice rejuvenated blood vessels in the brain, and consequently improved cerebral blood flow, while increasing neurogenesis and cognition. Scientists put forward the theory that, since calorie restriction and supplementation with young blood were effective in rejuvenating organs, they most likely have certain mechanisms in common. They therefore examined the molecule GDF11, which belongs to the GDF (Growth Differentiation Factor) protein family and is involved in embryonic development. GDF11 was already known to scientists for its ability to rejuvenate the aged brain. By injecting this molecule into aged mouse models, researchers noticed an increase in neurogenesis and blood vessel remodeling. The scientists also observed that the mice administered with GDF11 had lost weight without changing their appetite. This observation led them to believe that GDF11 could be a link between calorie restriction and the regenerating effects of young blood. The next step was to confirm this theory by studying adiponectin, a hormone secreted by adipose tissue which induces weight loss without affecting appetite. In animals that have undergone calorie restriction, the blood levels of this hormone are high. In animals that were administered GDF11, researchers also observed high levels of adiponectin, and this shows that GDF11 causes metabolic changes similar to those induced by calorie restriction. Until recently, there has been controversy over the role of GDF11 in aging, and its mechanisms were largely unknown. The findings of this study show that by inducing phenomena similar to those reported for calorie restriction leading to the stimulation of adiponectin and neurogenesis, GDF11 contributes to the birth of new neurons in the brain. Systemic GDF11 stimulates the secretion of adiponectin and induces a calorie restriction-like phenotype in aged mice Here, we present evidence that GDF11 induces a healthy calorie restriction-like phenotype together with brain rejuvenation in aged mice, and it acts by stimulating the secretion of adiponectin directly on adipocytes. We demonstrate a potent role for GDF11 as a metabolic actor in the aged organism based on the following findings: (a) systemic administration of GDF11 induced healthy weight loss as early as 1 week after treatment, (b) this weight loss reached a plateau throughout the rest of the treatment and was maintained for 3 weeks beyond the end of the treatment, (c) GDF11 levels were increased in aged mice that were subjected to calorie restriction, (d) metabolic changes were independent of GDF15 activation or anorexia, but correlated with changes in adiponectin levels and the insulin/IGF-1 metabolic pathway, (e) GDF11 activated adiponectin secretion directly from adipocytes, and (f) all the above changes correlated with a brain rejuvenation phenotype in aged mice. Cardiomyocytes Expressing SOX10 are Vital to Zebrafish Heart Regeneration A few higher animal species, such as salamanders and zebrafish, are capable of regeneration of limbs and internal organs, regrowing lost and injured tissue without scarring or loss of function. Numerous research groups are engaged in investigating the biochemistry of proficient regeneration, attempting to find the specific differences between species that might explain how it happens and why adult mammals are largely incapable of such feats of regrowth. Today's open access research is an example of the type, in which the authors narrow down on a specific cell population that appear in zebrafish hearts during regeneration, but not in human tissues. It may be the case that the mechanisms and capacity for adult regeneration do still exist in mammals, but are suppressed in some way, as suggested by the fact that the human ARF gene can shut down zebrafish regeneration. After all, we all managed to undertake the process of growing organ tissue during embryonic development. Alternatively perhaps a single crucial part of the adult regeneration mechanisms was lost over evolutionary time, and thus there is an opportunity to reinsert it into mammalian tissues via gene therapy or some other form of modern biotechnology. It still remains to be seen as to whether there are simple paths towards enabling greater adult mammalian regeneration, or, as seems equally likely, the situation is a complex mess that will take decades to decipher, and offers no easy path to therapy. Special cells contribute to regenerate the heart in Zebrafish In mammals, including humans, the heart muscle has a very limited capacity to recover after injury. After an acute myocardial infarction, millions of cardiac muscle cells, named cardiomyocytes, die, and are replaced by a scar. Unlike mammals, other vertebrates can recover much better from a cardiac damage. This is the case of some fish, including the zebrafish, a well-established animal model in biomedical research which shares with humans most of its genes. Zebrafish are extremely well suited to study organ regeneration. After heart injury, zebrafish cardiomyocytes can divide and the scar is replaced by new cardiac muscle. Now the researchers show that not all cardiomyocytes in the zebrafish heart contribute equally to regenerate the lost muscle, but that there is a specific subset of cardiomyocytes with enhanced regenerative capacity. A small subset of cardiomyocytes in the zebrafish heart, marked by sox10 gene expression, expanded more than the rest of myocardial cells in response to injury. These cells differed from the rest of the myocardium also in their gene expression profile, suggesting that they represented a particular cell subset. Furthermore, experimental erasure of this small cell population, impaired heart regeneration. The researchers want to find out whether the absence of such a sox10 cell population in mammals could explain why their heart does not regenerate well. If this is the case, the researchers believe that this finding could be
1.866707
openbmb/Ultra-FineWeb
Victorian and Edwardian interiors are often thought of as one and the same. In reality, once this era had gotten underway, the differences in style and taste began to deviate massively from those popular during the reign of Queen Victoria. Here's a brief overview of the important elements to consider if you're styling an Edwardian themed home: Victorian rooms were often cluttered with ornaments and decorated in very dark colours and patterns. The Edwardian age marked the beginning of a fresher, airier interior designs that reflected the positivity of the period. The era that saw the birth of powered flight and the sinking of the Titanic was dominated by the tail end of the "Arts and Crafts" movement, seeing a heavy use of Medieval and natural designs that showed an anti-industrial theme harking back to Britain's rural past. Art Nouveau added to this style, with natural forms and structures, flowers and plants, and curved lines. The Main Elements of the Room Edwardian rooms mainly featured highly polished flooring with rugs instead of carpets. Oriental themes were very popular in this period, and for high traffic areas and kitchens, red tiles were the order of the day. When it came to colour, the Edwardians favoured pastel colour schemes and floral designs: Primrose, leafy greens, lilac and even subtle greys. Wallpapers too, should reflect this light and airy feel, and feature curvy, floral shapes and pastel hues. Art Nouveau styles of furniture were often mass produced reproductions, and proponents of the style had little problem with this, leaving you open to a lot more choices than the equivalent Victorian home. Wing chairs, rococo, baroque and empire styles are all appropriate, although clutter is definitely not. This was also the era of wicker furniture. It's easy to track this down, so if you see anything appropriate in a junk shop, snap it up! The Dawn of Electric Lighting Electric lighting wouldn't necessarily have been standard throughout the middle classes, but the larger estates and affluent city houses would definitely have been keen to electrify their homes. You can can show this influence by including lamps with period shades, such as those with tassels and soft fabric surrounds. Tiffany lamps, as well as large, decorative ceiling lights were also very fashionable, especially with glass etched shades. It's important not to overdo an Edwardian themed room when it comes to decoration. Choose some key pieces and go sparingly. William Edwards, makers of fine bone china, recommend an Edwardian china set on display: "Royal Worcester china and similarly styled reproductions would make a very elegant and fitting ornamental display for a dining room or even a living room". Other accessories include old gramophones with conical horns, large cake stands and silver photo frames in art nouveau styles. Flower arranging was a popular Edwardian pastime too! Decorating your home using Edwardian design and ornaments is a fantastic alternative to the more cluttered, and often more expensive Victorian styles, with its airy colours, floral wallpapers and art nouveau ornaments, your home will have a unique period look all of its own! By Harry Price – Harry Price is a writer and personal trainer. He left his career as an interior decorator to pursue his own business ventures and has never looked back.
1.323737
m-a-p/FineFineWeb
This article describes a generalized approach to measuring subjective notions of quality. It shows how using a particular mathematical framework can yield several beneficial properties. These properties allow for measurements of complex, subjective notions of value or quality that are intuitive and easily tailored to a particular individual and data set. This example shows how the selection of different metrics can facilitate the incorporation of desirable features of the final scoring methodology. What's in a number? In my continuing saga on metrics, I quoted one IEEE article about the goal — "the main objective is to develop measures that are consistent with the subjective evaluation." I gave an example using litigation witness files to describe what this might look like. I was focusing on showing how a "QE" (Quality-Efficiency) score encourages competition of desirable characteristics of a given legal service. In this post, I'm examining some basics of how to select numerical formulas that correlate with our subjective notion of quality. This post is somewhat technical (mathematical) and geared toward legal technologists. My goal, however, is that difficult parts can be glossed over and the overall points will still be understandable. As a disclaimer, I have no idea how much of this might have been done elsewhere (some of it dates back to my dissertation). Also, although I use a legal perspective on the scoring, nothing in the approach is specific to law. Example: selecting an attorney Selecting an attorney is in many ways similar to selecting an online date. We care about several factors and have to combine them to get a ranked list of candidates. Matching algorithms are similar to how we measure, say, general search results. In search, we may want to factor in topic, author, date, or even geographic region (just because we'd like to incorporate certain factors doesn't mean that the system is able to accommodate them). Similarly, as we consider measuring something like QE, we have to combine several factors to come up with a selection criteria — which provider should we use? How we incorporate factors is an individualized choice. A good result for one person isn't necessarily a good result for someone else. Moreover, what matters to us is impacted by what's available. If we don't find what we're looking for (search results, vendors, online dates, whatever), we have to adjust our criteria. We're looking for the best from a pool of potential candidates, and that pool may vary from region to region or from day to day. The example of selecting an attorney highlights how we combine characteristics in general. Step one: combining normalized scores Suppose I want to select an attorney to help me with a pressing legal issue. The most simple case would be one in which I care only about a single factor — say cost. Suppose I have a set of attorneys who's only distinguishable difference is how much they charge for handling my issue. I can just rank them based on who is the least expensive. Of course, I'm likely to care about other factors too. Suppose I also care about distance. If I cared only about distance, I could do the same with distance as I did with cost and just rank the attorneys by how close they are. But I need to combine cost and distance to get a single ranked list (i.e. reducing two dimensions down to one) — the order in which I'll look through the results in more detail to make a final selection. Suppose, for Case 1, that one attorney would charge $1000 and is 20 miles away, and that another attorney would charge $2000 and is 10 miles away. I might prefer driving the extra 10 miles to save $1000. I could try adding the numbers — 1020 vs. 2010 — and take the smaller. But what if, Case 2, the less expensive attorney is 100 miles away and I don't want to drive that far? The numbers don't work: 1100 ($1000, 100 miles away) is worse than 2010 ($2000, 10 miles away). Another approach might be to multiply cost and distance. For Case 1, I'd get a score of 20,000 for both attorneys (1000 * 20, and 2000 * 10), even though I prefer the less expensive one. This puts each mile of distance equal to $100 in fees, which doesn't fit what I want. Case 2 seems better. Even though the attorney that is too far away would get a score of 100,000 (1000 * 100), an attorney who is 21 miles away, an acceptable distance for the reduced cost, would get a higher score than the more expensive one, which also doesn't fit what I want. To reconcile this, we can take each factor and derive a "normalized" score — a range from 0 to 1 — such that 0 means completely unacceptable and 1 means perfect. We then multiply the normalized values to arrive at a new score whose range is also 0 to 1, where 0 means completely unacceptable and 1 means perfect. To get the values for each factor, we could define ranges (there are many ways to accomplish this). Suppose that for distance, we define three levels of distance: very close, close enough, and too far. Very close, anything within, say, 10 miles, will get a 1. An attorney who is farther than, say, 50 miles, is just too far away, so we'll give that a 0. Any attorney between 10 and 50 miles will get a 0.5. Similarly, we can stratify cost: 1 for $0 to $999, 0.75 for $1000 to $1999, 0.5 for $2000 to $2999, 0.25 for $3000 to $3999, and 0 for $4000 and above (just too expensive). Case 1 gives a score of 0.375 (0.75 * 0.5) for the less expensive attorney, and 0.25 (0.5 * 0.5) for the more expensive one. For this type of scoring, the larger number is better, so we would select the less expensive attorney (who happens to be within our acceptable driving distance). Case 2, however, gives a score for the less expensive but very far attorney of 0 (0.75 * 0), and we would select the more expensive attorney — the only one within our distance limits. These values are closer to what we wanted. At this point, we have a ranked list based on a "cost-distance" score with some intuitive characteristics. First and foremost, we get an ordered ranking that correlates to our intuitive notion of how we want to balance the trade-offs between various criteria (e.g. cost and distance). (This is a "non-strict total order", since if one lawyer is farther but less expensive, and another lawyer is closer but more expensive, the two might get the same overall score.) That's because we're effectively assigning an equivalency ratio between the various characteristics (e.g. a certain change in distance is equivalent to a certain change in cost, though not necessarily linearly). Second, the scoring approach needs to be heterogeneous — it has to be able to accommodate and combine any type of data. So far, I described distance and cost, which are already linear data types. The scoring mechanics should be language independent as well. Below, I'll show an example of extending this to other types of (nonlinear) data. Third, we have a filter that excludes all unacceptable candidates. In this case, we exclude any attorney that is either too far or too expensive. We get this property by multiplying values between 0 and 1, where 0 is unacceptable. If a candidate is unacceptable in one of its characteristics, the value of that characteristic is assigned 0. Since we multiply the values, and anything times 0 is 0, unacceptability in any characteristic sets the final score to 0 and thus excludes it from the candidate set. This is an important property. If we have any hard limits for any characteristic (too far, too expensive, etc.), then it simply doesn't matter if a candidate fits all the other characteristics perfectly. The flip side of this, however, is to use filtering judiciously. If we set a hard limit at, say, 50 miles, but an otherwise perfect lawyer is 51 miles away, did we really want to exclude her? Furthermore, since we need to adjust our scoring based on the candidate set, we may well need to set our hard stop differently on a case-by-case basis. If we're looking for a lawyer in an urban area, we might be able to justify a closer hard stop on distance than if we're in a rural area — and perhaps the inverse on cost for the same reason (density of lawyers). Fourth, I'm incorporating normalized scoring (such as relevance in Information Retrieval literature) — all characteristics are assigned a range of values from 0 to 1. This is an extension of boolean scoring, which is either 0 (exclude) or 1 (include). Boolean scoring also has the filtering property. However, boolean scoring doesn't have the ranking property — all acceptable candidates get the same score: 1. Normalized scoring is a strict generalization of boolean scoring, since any characteristic can be treated as a boolean and the multiplicative combination of scores will still function properly. Fifth, by using normalized scoring, I'm able to apply a hierarchy property. Hierarchical scoring allows us to derive sub-scores and treat them as atomic scores, or take an atomic score and decompose it into smaller elements without impacting the way that characteristic is used. As an example, in my prior post
1.511298
openbmb/Ultra-FineWeb
Name of Activity New Test Activity Author Tufts STOMP Keywords keyword 1, keyword 2, keyword 3, keyword 4, keyword 5, 1 Hour Total, NXTs, LEGO, k-3 Subject NXTs Grade Level K, 1, 2, 3 Time 1 Hour Total Brief Description Brief description goes here. Describe what the activity consists of in 1-4 sentences. Lesson Objectives: - To teach students about _____________________. - To introduce ____________________ vocabulary words. - To familiarize students with LEGO building. - To introduce programming to students grades k-3. Materials Needed: - Material 1. - Material 2. - Material 3. - Material 4. - Material 5. - Material 6. - Worksheets 1-4. Preparation and Set Up: - Describe all set up here. - Hand out worksheets. - Split students into groups. - etc etc etc. Necessary Background Fill in necessary teacher background here. This can include vocabulary words, general information, or any prepared information about important subject matter here. 1. Step by step procedure goes here. 2. Include the Engineering Design Process wherever applicable. 3. continue all the way through the activity, start to finish. Extensions: Please elaborate here if there is any way to make the activity more difficult for students who finish the activity early or are more experienced than other students in the subject matter. Modifications: Please elaborate here if there are ways to change this activity to make it slightly different or better in any way after your experience with it. Tagged with → Share → Leave a Reply Switch to our mobile site
2.389853
Zyphra/Zyda-2
That would fit a context such as KdrV A224/B271, where Kant equates [den Begriff eines Triangels] konstruieren (to construct/construe the concept of a triangle) with to give [the concept] an object, i.e. to give it what he elsewhere calls a reference [Beziehung] to objects, and hence a significance [Bedeutung] (A1456/B185), or, taking Bedeutung more broadly: to give it an imagined significance [eingebildete Bedeutung] (A84/B117), as opposed to a mere logical significance (A147/B186). To construe, that is, interpret, is to give significance. Unfortunately, this appears not to be an accepted sense of the German verb (and moreover, according to the OED, is only a late and derivative sense of the English one). Moreover, we have to take into account another use of konstruieren in the exact same passage (from the proof of the First Postulate). Kant also talks there about the synthesis wodurch wir in der Einbildungskraft einen Triangle konstruieren (A224/B271). Whatever that means, it doesnt mean by which we construe (= interpret) a triangle in the imagination. Many would take it, instead, to mean something like: by which we construct (= build up) a triangle out of imaginary pieces, using an imaginary compass and an imaginary ruler, on a piece of imaginary paper. But that in itself makes little sense. Moreover, to the extent that it does make sense, it leads straight back to Descartess argument that we cant do any such imaginary drawing in the case of a chiliagon (a point of which Kant is well aware: see A140/B17980). Finally, such an understanding of konstruieren wouldnt work at all in the first example, where Kant talks of constructing the concept of a triangle. A concept presumably cant be drawn on any paper, no matter how imaginary. So my new thought is this: that both of these ways of talking are short for: einen Begriff mit einem Gegenstand konstruieren, to construe a concept with an object (on the model of: to construe a verb with an object). If we can construe a concept with an object (via a schema), then we know that it is at least the concept of an ens imaginarium (not, like the concept diangle, that of a nihil negativum [A291/B348]; and not, of course, of an ens rationis or nihil privativum, either). What hes saying in the proof of the First Postulate (i.e., the proof that what agrees with the formal conditions of experience (according to intuitions and concepts) is possible [A218/B265]) is that to promote a concept from the concept of an ens imaginarium (which, although not impossible, is still a kind of nothing) to the concept of a possible being, we need to connect the procedure of the imagination in the schema of that concept to the synthesis of the imagination in perception. So the object of the concept triangle would remain always only a product of the imagination if it were not that The very same [eben dieselbe] imaging synthesis by which we construe a triangle [with its concept] in the imagination is wholly the same as [mit einerlei] that which we exercise in the apprehension of an appearance, in order to make from it an empirical concept. (again, A224/B271) The concept triangle passes this test, but the concepts of pure time and pure space as such, die zwar Etwas sind, als Formen anzuschauen, aber selbst keine Gegenstände sind, die angeschaut werden [A291/B347], do not. The synthesis by which the pure imagination produces time and space as pure images of the categories is not the same as any synthesis of the empirical imagination in perception (but is rather a condition of possibility of all such empirical synthesis). Hence the concepts of time and space are concepts of entia imaginaria.1 However, Im not sure yet if this will work in other places or whether Kant anywhere actually uses the locution to construe a concept with an object. 1 The ursprünglich pure intuitions of time and space, I take it, do not have objects: intuitions without concepts are blind.
2.063285
HuggingFaceFW/fineweb-edu
Tag Archives: Latvia Latvia 2016: The Gathering Storm The weakening of the West and its impact on Latvia Latvian inner political and economic weakening The challenge of 2017-2018 Comments Off on Latvia 2016: The Gathering Storm Filed under Current Events Latvia and Ukrainian People's Republic 1920 map of claimed lands by Ukrainian People's Republic. Note: Eastern Ukraine – Galicia and Lviv has been already to Poland. Ukrainian People's Republic (Українська Народна Республіка УНР or UNR) was first modern Ukrainian national statehood that existed between 1917 and 1921. Similarly to Belarusian People's Republic (BNR) it did not survive the tides of war, however contrary to BNR, UNR received greater support and recognition from Latvia. Latvian officers also took part in UNR national forces and both sides had high hopes towards each other. On 1920 as Ukraine was divided between Soviet Russia and Poland in pursue for peace with Soviets, Latvia had to abandon its support for UNR. Article highlights Latvian – Ukrainian diplomatic relations and Latvian participation in UNR armed forces. On March 17 1917 after the collapse of Russian Empire Central Council of Ukraine (Українська Центральна Рада UCR) was established in Kyiv. One of its main demands was national autonomy that was not supported by Russian Provisional Government in Petrograd. In response Ukrainian national forces under the command of General Pavlo Skoropadskyi started to assemble to defend Ukrainian sovereignty. On October 25 (November 7) Bolsheviks seized power in Petrograd and their supporters took over some areas in Ukraine. UCR managed to control much of Ukraine and on November 7 (November 20) the Ukrainian People's Republic was proclaimed. Meanwhile in Kharkiv on December 11-12 1917 Ukrainian Soviet Republic was founded. Bolsheviks advanced and gained control over much of Ukraine and on February 8 captured Kyiv. Chief Otaman Symon Petliura who commanded UCR armed forces faced gruesome defeat at the battle of Kruty where Ukrainian 1st Student company and Cadet Corps suffered great casualties. On January 9 (22) UNR again proclaimed full independence and severed all ties with Russia. On February 9 in Brestlitovsk UNR signed peace treaty with Germany and Austria-Hungary gaining their military support in return for food provisions. German forces entered Ukraine and on March 1 captured Kyiv. On March 3 Soviet Russia ceded Ukraine to Germany in peace agreement in Brestlitovsk. UNR forces of 15 000 men entered Kyiv and Crimea. UNR was recognized by Germany, Bulgaria, Austria-Hungary, Turkey and Great Britain. Ukraine was taken over by Germans and Austrians who came in early conflict with UCR who protested against German military courts. German authorities formally dissolved UCR. In response UCR proclaimed new leftist leaning UNR constitution. Ukrainian right-wing rushed to prevent Ukrainian-German confrontation and rise of left-wing on April 29 seized power. General Pavlo Skoropadskyi became dictator under the title of Hetman of Ukraine. UNR faced resistance from Bolsheviks and peasants lead by anarchist Nestor Makhno. More countries however recognized UNR such as Finland, Poland, Romania, Switzerland, Denmark, Netherlands, Norway and Persia. As of November 11 1918 when Germany and Austria-Hungary collapsed situation changed in Ukraine. UNR elites supporting Allied powers created Directory lead by Symon Petliura, who in December 14 deposed hetman Skoropadskyi and proclaimed second Ukrainian People's Republic (during hetman's dictatorship UNR was called "Ukrainian State"). Meanwhile in Western Part of Ukraine a Western Ukrainian People's Republic on October 19 1918 was proclaimed in attempts to split from AustriaHungary who controlled Lviv (Lemberg). On January 1919 it united with UNR. The Western Ukraine with Lviv in center was claimed by Poland and both sides engaged each other in battles until June 1919. Soviet Russia immediately after German surrender to Allies canceled Brestlitovsk peace agreement and UNR had to declare war on them on January 16 1919. On February 5 Bolsheviks again captured Kyiv and reached Zbruch River west of Ukraine and entered Crimea. Ukrainian peasants and anarchist groups resisted Bolsheviks while un summer of 1919 the monarchist White Guard South Russian Volunteer army lead by general Anton Denikin attacked Bolsheviks and captured Kharkiv on the way to Moscow and on August 31 entered Kyiv. As his forces were defeated and chased away Bolsheviks again took over much of Ukraine in March 1920. Situation demanded an alliance with Poland that was reached by Symon Petliura. On April 26 1920 Polish-Ukrainian forces attacked Bolsheviks on May 7 captured Kyiv. Polish forces reached river Dnipro. Bolsheviks ignited counter offensive and On June 12 Poles abandoned Kyiv, on July 4 Bolsheviks started attack in Belarus and reached Warsaw. On August 12-17 the Bolshevik offensive was finally stopped at the gates of Warsaw. On March 18 1920 in Riga Poland and Soviet Russia signed peace agreement dividing Ukraine in two. Central, South and East Ukraine was granted to Soviets while Western Part of Ukraine including Lviv (Lwov in Polish) was ceded to Poland. Recognized also by Ukrainian delegation the Riga peace agreement was death sentence to UNR. Symon Petliura dismissed Directory and withdraw his forces to Poland where he was interned. Rumania and Czechoslovakia also gained Ukrainian ethnic lands. Symon Petliura lead the UNR in exile until he was assassinated by Soviet agent on 1926. UNR continued to work in exile in Poland until 1939 and the moved to France where it ceased to exist after Nazi occupation. After the war in western exile Ukrainian National Council (Українська Національна Рада) that existed until 1991 when it recognized new Republic of Ukraine that formed in result of collapse of the Soviet Union. During dramatic and fast changing events in Ukraine during Soviet-Ukrainian war many Latvians were involved army in state matters. Firs before the First World War and during the war some Latvians traveled to live and work in Ukrainian provinces of the Russian empire and secondly the Latvian officers who served the Russian army were sent on duty there. Large numbers of Latvians ended up in Ukraine as refugees during _PHONE_. Also campaigners for Latvian independence were interested in Ukrainian independence movement and were seeking for cooperation. On September 8-15 (21-28) in Kyiv the UCR organized "congress of the minor nations" where 80 representatives took part along with 10 from Latvia. Latvians were represented by Zigfrīds Anna Meierovics from Vidzeme land council Kristaps Bahmanis from Kurzeme Land council and Staņislavs Kambala from Latgale land council. Z. A. Meierovics gave speech describing the history of Latvian history and present situation and voiced call for Latvian self-determination. Meierovics, Kambala, Bahmanis and lecturer at Kyiv University M Bruninieks were elected in All-Russia Council of Nations. On November 18 1917 in Valka the newly founded Latvian National Council (LNC) also discussed Ukraine and judged that Ukrainian politicians are generally friendly towards Latvians like rest of the small nations. On December the Council received telegram from UNR about their declaration of independence and replied back with warm greetings towards Ukraine and voiced support for Ukrainian freedom and federation of nations. On January 1918 LNC decided to send representatives to Ukraine to gain support promised by UCR. At first Latvians wanted to travel to Brestlitovsk to take part in Ukrainian-German peace talks to gain things to their favor, however the idea was dropped and Latvian representative K Bahmanis went to Kyiv on February. There he contributed to the creation of Kyiv Latvian Central Committee and spread information about the work of LNC. Because of war activities he could not contact Latvia and his report about his activities was only reviewed in June. Bahmanis became the representative of the Latvian Provisional Government in Kyiv and since 1919 visited new governments in Georgia, Armenia, White Guard Armies in Crimea, Don and Kuban. He returned to Latvia in September 1920. Latvian General Pēteris Radziņš who served in UNR amy Many Latvians who were at Ukraine decided to join Ukrainian national armed forces lead by General Pavlo Skoropadskyi. Latvian officers also served in Symon Petliura Directory army. Most known was colonel Pēteris Radziņš who was chief of organizational and training department of the General Staff. After hetman was deposed he served as deputy to the chief of General Staff Mykola Yunakiv. On September 1919 escaping UNR defeats against White Guards and Bolsheviks he got himself in Poland and then returned to Latvia. There he was appointed for the Chief of the Commander-in-Chief Staff of the Latvian army. From 1924 to 1928 he was commander-in-chief of the Latvian Army. He also was author of many military history books and died in age of 50 in 1930. Lieutenant Colonel
1.463966
Zyphra/Zyda-2
Researchers create interlocked self-assembled rings that can be seen under a microscope An international team of researchers, including Keele's Dr Martin Hollamby and led by Professor Shiki Yagai of Chiba University in Japan, has succeeded in forming and studying self-assembled interlocked rings called "nano-poly[n]catanenes" that contain just one molecular ingredient. In 2016, the Nobel Prize in Chemistry was shared by Jean-Pierre Sauvage for the design and synthesis of molecular machines, after he managed to interlock two ring-shaped molecules into what is called a "catenane". Unlike ordinary chemical bonds, the molecules in catenanes are linked like a chain, and so the links can move relative to each other. The synthesis and characterisation of such structures are notoriously difficult, particularly when the component rings are not held together by strong covalent bonds. This work, published in Nature, reports the first template-free synthesis of nano-poly[n]catenanes via molecular self-assembly. The component rings ("nanotoroids", 30 nanometres) contain around 600 identical small molecules. By altering self-assembly conditions, the group were able to create intricate structures, including a nano-catenane with interlocked rings in a linear arrangement, which has been named "nanolympiadane" in homage to the catenane system "olympiadane" first reported by Fraser Stoddart and colleagues in 1994, and the well-known symbol of the Olympic games. The group designed methods to purify the rings, removing any material that hadn't assembled as desired, and found that the addition of such rings to the hot monomer solution facilitates the formation of new assemblies on the surface of the rings, a process known as secondary nucleation. Informed by the step-wise nature of the formation mechanism, they added monomers sequentially, and were thereby able to create poly[n]catenanes with up to 22 rings. Given the scale of these structures (already hundreds of nanometres), it is likely that only minor further refinement of the sequential addition process may build structures that are sufficiently large for purification by simple filtration. This would allow for in-depth study of the unique physical properties that a structure made up of miniscule interlocked chain links may have, and their potential for creating molecular machines. Dr Hollamby said: "Substantial interlocked structures of molecules have been made before, but the rings that make them up have previously been held together by strong covalent bonds. This report marks the first time that this has been done at this scale, and with such a fine level of control, for structures that are built up from a single small molecule interacting with itself. "Keele's contribution lay in the analysis of small-angle scattering collected at ISIS Neutron and Muon Source and Diamond Light Source which alongside multiscale molecular simulations by the group of Prof. Pavan (Polito, Italy & SUPSI, Switzerland) were instrumental in characterising the structures, identifying where the secondary nucleation took place, and the mechanism behind it." Lead author Professor Shiki Yagai added: "The innovative finding of this research lies in the utilisation of the self-assembly characteristic of the molecules. We were able to create intricate geometric mesoscale structures without using complex synthetic methods." "This paves the way to create even more complex geometric compounds such as 'rotaxane' and 'trefoil knot' at a similar scale. As the molecular assemblies used in this research are made of molecules which react to light and electricity, they can potentially be applied to organic electronics and other molecular machines." - Plans unveiled for exciting Keele in Town development - Keele celebrates strong links with NHS to mark service's 75th anniversary - Celebrating our Good and Outstanding Ofsted results - New course launched at Keele to prepare students for green industrial revolution - Keele welcomes 10,000 guests to campus for summer graduation week
2.312469
openbmb/Ultra-FineWeb
reducing inflammation, swelling, and headaches; and improving healing in general. The following is a deeper synopsis of some of the health benefits associated with a lymphatic drainage. 1. Promotes skin care: Draining the lymphatic system will improve your complexion and help reduce wrinkles, fine lines, and puffiness around your eyes. It can also reduce old scar tissue and minimize new scars before and after cosmetic surgery. 2. Reduces headaches: Most headaches have a component of lymphatic congestion, and lymphatic drainage can potentially help reduce pain and discomfort and improve blood flow. 3. Decreases swelling: Fluid often builds up in the tissues after long periods of travel and immobility, and this leads to swelling and tenderness. For instance, arthritis is a condition where the joints are congested with fluid. A lymphatic drainage will improve conditions where swelling is an issue. 4. Benefits overall healing: After an injury or surgery, tissue may be sore or swollen. However, lymphatic drainage can reduce inflammation, gently drain tissue, and improve overall healing. 5. During and after pregnancy: Fluid retention is common in pregnant women, and lymphatic drainage can provide relief in the feet and legs. Lymphatic drainage can also enhance breastfeeding, especially when breasts are sore and ducts are blocked. Final Thoughts on Draining Your Lymphatic System The lymphatic system is designed to eliminate waste from your body. Most people notice great improvement after draining the lymphatic system. This is because it benefits the skin, reduces swelling and pain, and improves circulation and the flow of lymph and blood. In this article, we explained how to cleanse your lymphatic system naturally with diet, massage, exercise, dry skin brushing, hydration, yoga, deep breathing, herbal remedies and supplements, homeopathic remedies, juice cleanses, acupuncture and acupressure, and soaking in a bath or sauna. The rejuvenating effects of your lymphatic drainage could have you looking and feeling great as a result. - Home remedies for swollen lymph nodes - How to treat swollen lymph nodes in neck: Causes, symptoms, and massage tips - Swollen occipital lymph nodes: Causes, symptoms, and treatments - Swollen lymph nodes under chin: Causes and natural treatment tips - Swollen lymph nodes in armpits: Common causes, treatment, and home remedies - The Benefits of Lymphatic Drainage Massage Article Sources (+) MacGill, M., "What does the lymphatic system do?" Medical News Today, February 23, 2018; _URL_ last accessed June 29, 2018 . Ekici, G., et al., "Comparison of manual lymph drainage therapy and connective tissue massage in women with fibromyalgia: a randomized controlled trial," Journal of Manipulative and Physiological Therapeutics, Feb. 2009; 32(2): 127-133, doi: 10.1016/j.jmpt.2008.12.001.
2.021754
m-a-p/FineFineWeb
Wiring The BackcountryBack in the early 1920s, what was then called the Idaho Primitive Area got telephone communication. The descendants of one such worker explore his journal. By Betty Derig King HillA place of rocks and steep grades was in need of irrigation if it were to become habitable. Here's a story of amazing engineering feats to make that happen. By Dean Worbois The King Hill DiariesAs an accompaniment to this month's Spotlight City feature, a researcher of pioneer diaries provides excerpts from the writings of early travelers in the area. By Jerry Eichhorst Here Comes the NightBy Jon Mills Riverboat, AhoyBy Grove Koger On the BeachBy Michael Stubbs Snake River Plain With Bat Man and Snake LadyBy Kris Millgate After SummerBy Rachel Gattuso Recycling ForeverBy Les Tanner Hooked on BooksBy Joe Davis Idaho at Large
1.043143
m-a-p/FineFineWeb
AnoxKaldnes Z-MBBR™ AnoxKaldnes™ Z-MBBR Z-MBBR by AnoxKaldnes, biofilm in 3D. << Back to AnoxKaldnes website A different approach to carrier development Traditionally, carrier development has aimed for larger protected surface area, but has allowed for a more or less uncontrolled growth of the biofilm. This lack of control of biofilm thickness is a major uncertainty for process design based on carrier surface area, and also limits the applications for which the MBBR can be a competitive solution. Z-MBBR - The new dimension The new carrier design is covered with a grid of defined height, allowing the biofilm to grow on the outside of the carrier in a protected environ-ment. Due to scouring from other carriers, the biofilm cannot grow higher than the grid height, and so biofilm thickness is controlled. Applications of Z-MBBR As all process limitations related to uncontrolled growth of biofilm and clogging of carriers are eliminated with the Z-carriers, they are especially useful in applications where the biofilm tends to grow thick, such as nitrification at higher organic loads. Z-carriers with different biofilm thicknesses have been evaluated for different applications, with very interesting results. As for example, extremely thin biofilms have shown to open for a new, more efficient application of mainstream ANITA Mox.
1.12271
Zyphra/Zyda-2
Learn about flying kites, hot air balloons, remote control airplanes, gliders, and space shuttles. The fascination of flying an airplane, launching a rocket, conquering space and becoming an astronaut or pilot will come alive as you explore aviation and space. Or explore the science of rocketry. Use the knowledge you gain to design, build, and launch your own two-liter pop bottle rocket. If you are already signed up for this 4-H project, come on in and get started. If you haven't signed up yet, contact your county 4-H office and become a member today. Skills to Achieve Mastery - Develop science and technology literacy as it relates to aerospace. - Improve communication and teamwork skills by working on group activities. - Develop responsible attitudes about science and how science relates to the real world. - Gain experience in problem solving and decision making using science process skills. - Gain experience in SET abilities. Back to 4-H Project Home
1.97119
m-a-p/FineFineWeb
Race Engine Airflow - Where Air Goes...Or Should Written by on September 1, 2003 You can't see air. But there's a way to examine flow paths in the intake track and combustion space, and it's not "snake oil." There's power in the method. Race Engine Airflow - Where Air Goes...Or Should View All 8 Photos Consider this: Engines don't burn just fuel, they burn fuel and air. The more air, the more power, and fuel delivery is adjusted accordingly...or so the axiom goes. In reality, air quantity must be equated with air quality, for without proper conditioning of mixtures, combustion efficiency will suffer. More air isn't always more power...as you will soon discover. Ctrp 0309 01+race Engine Airflow 2/8 A simple dye spray may be enough to unlock more power in a racing engine. Some Fundamental Considerations Let's say you have an engine that produces a given level of power, demonstrates certain brake specific fuel consumption (BSFC), and requires a known quantity of spark timing. Further, assume this engine to be our "baseline" for discussion. As you might assume, these conditions are produced with what we'll also call "baseline combustion efficiency." Now, without changing net airflow into the engine, suppose we perform a set of modifications intended only to improve mixture quality (homogeneity). In conventional terms, this simply means keeping more fuel in suspension prior to and during combustion. We want to emphasize that net fuel and airflow are not changed, only mixture quality. Ctrp 0309 02+race Engine Airflow 3/8 Short bursts of aerosol machinist dye work well to establish airflow patterns. Excessive use can mask the contrast of how the dye represents the intensity of where separated fuel will migrate. Before concluding inlet port and chamber designs, effects of the intake manifold should be determined. While this does not include the influence of a carburetor, it does reveal the impact of ports and runners working together. Note proximity of can to entry, as demonstrated by Wells Racing's Ricky Stapleton. From these types of modifications, some interesting results can occur. First, improved mixing (homogenization) tends to reduce fuel particle size. The smaller their size, the faster the burn, generally speaking. Then, as stated in previous technical materials in Circle Track, cylinder pressure rises more quickly, enabling slight reductions in total ignition spark timing. As either or both of these conditions develop, net torque on the crankshaft increases, largely by the reduction in negative work (pressure on the piston during combustion and prior to TDC compression position) and a quicker burn. These conditions trace directly to improved mixture quality, which then becomes the enabler for torque increases. More to the point, it is the "construction" of improved air quality that supports more efficient fuel atomization and burn efficiency gains. The Problem of Air/Fuel Separation This is a classical problem when changing the direction or flow rate (or both) of fuel-suspended airflow. The relative differences in compressibility and weight (by volume) of air and fuel leads to their departure during changes in energy state (flow direction, rate, or both). It is the condition of the air that tends to affect how fuel is delivered. As an analogy, the air steam could be equated to a flowing river and fuel likened to leaves thrown into the stream. The motion of the river stream significantly affects the direction and movement of the leaves. By affecting the characteristics of the stream, movement of the leaves can be directed. Such is the case when transmitting fuel particles through an engine's intake track. Ctrp 0309 03+race Engine Airflow 4/8 In this LS1 chamber, note the wash of fuel against the bowl area and across the spark plug boss. While this direction of mixture motion toward the exhaust side of the chamber is desirable, the heavy concentration of dye beneath the valve seat and absence around the backside of the chamber wall is not. Interestingly, many cylinder head modifiers spend inordinate amounts of time sorting out port designs, valve seat and bowl dimensions, and combustion chamber shapes to optimize net airflow. Not to diminish the value of this approach to increasing an engine's output, especially if fuel is admitted up-stream of the combustion space, but omission of how such airflow affects the delivery of fuel can lead to an imbalance between increased air and power. Initially, the idea of using some form of atomized spray that would leave its mark on flow surfaces was conceived to address the air/fuel separation problem. In time, as the windows of information about this approach began to open, it became evident that other aspects of improving combustion efficiency could be explored. In fact, this sequence of investigation led to the notion of using kinetic energy already contained in the inlet air stream to aid re-suspension of separated fuel by the method of dimpling or a comparable approach to surface imperfections. There are, of course, more sophisticated methods to study active flow and the influence of variables that affect its behavior and characteristics. However, most of you don't have access to CFD (Computational Fluid Dynamics) or comparable techniques. So at the more grassroots level, the approach described in this story is worthwhile. Ctrp 0309 10+race Engine Airflow 5/8 While canting the back wall of this combustion chamber may have improved raw airflow (as measured on a bench), note the concentration of fuel wash and lack of movement in the area of the spark plug nose. Rotation of the flow in a counter-clockwise direction (shown) is favorable, but the collection of raw fuel just past the exhaust valve pocket is not. Matching such patterns with piston crown shapes and surface texture is important to boosting combustion efficiency. One of the purposes of this story is to reveal the process of using both dye traces and flow path surface roughness to aid fuel atomization and suspension in the air stream. Although considered by some engine builders to be a "crutch" or "band-aid" to resolving a problem inherent in virtually any internal combustion engine, the latter of these two methods is proven, recognized by the U.S. Patent office as viable intellectual property, and increasingly practiced wide-scale by builders of several types of engines. As previously stated, it is not "snake oil" and is worthy of the time taken to understand and practice, at least in the opinion of this writer. In the Real World of Making Power... Dennis Wells of Wells Racing Engines shared comments about his experience with the airflow/dye method of flow analysis. Ctrp 0309 09+race Engine Airflow 6/8 This is good. Note uniformity of pattern that includes activity in proximity of the spark plug nose and lack of "fuel" buildup both against the chamber's back wall and sidewall of the exhaust valve pocket. Uniformity of pattern suggests a narrow range of air/fuel ratios will be present in the combustion space, as opposed to a wider range of ratios when patterns indicate sharper contrast of dye residue. "Before I began practicing this method, we'd tear down an engine, look at the burn patterns on piston tops and combustion chambers and wonder about what we saw. Sometimes, we'd see no residue; other times, we'd see varying shades of carbon buildup. Even though it occurred to us these patterns were probably an indication of something we should understand, we really didn't. "Then, I was introduced to using a spray-type dye during airflow bench testing of cylinder heads. By studying areas where the dye was really concentrated, it appeared these locations indicated fuel wash or separation from the airflow. What really got my attention was when we started changing the surface texture of these areas and began to see power increases with little or no changes in total air or fuel flow. Exhaust gas temperatures also began to drop, suggesting the burn was taking place faster and that more of the fuel already in the combustion chamber was now being burned. We also discovered less spark timing was required for best power," Wells said. "As I've used this method of airflow study, we've found other areas for modifications that help increase horsepower. One of these has been in working with two-plane manifolds and the use of carburetor spacers. I never realized why spacers seemed to make more power until we did some dye work and discovered the problem of fuel separating on the manifold's upper floor. We discovered that as the carburetor is raised, this separation problem was reduced and more power was produced from about the same amount of fuel and air we used before adding the spacer. In some cases, we found it was possible to lean down the carburetor and make even more power. Sometimes this wasn't necessary because raising the carburetor also tends to reduce fuel-metering signal and makes the engine leaner, anyway. Ctrp 0309 08+race Engine Airflow 7/8 In areas where fuel is impacting combustion surfaces and "washing" into rich air/fuel ratio development, surfaces can be dimpled to create eddies or boundary layer activity intended to restore separated fuel into the air stream. While this approach may be considered a "band-aid" method, it can be highly effective, cost-effective, and power producing. "Overall, there are several things (construction techniques) we've found helpful using the dye method," Wells continued. "One is that it points to where flow surface texture changes will help. We like to use dimples in these areas because they tend to put separated fuel back into the air stream. What this does is allow otherwise wasted fuel to be burned. This is power, any way you look at it. We also discovered that how air is delivered to the combustion chamber, when measuring flow patterns with a head alone, usually changes when you bolt on the intake manifold. So
1.571733
Zyphra/Zyda-2
Organic Planting With Cactus Soil Cactus soil is popular for many types of potted plants. It is rich in nutrients, does not retain excess moisture and has an acidity level that falls right in the middle of what is considered optimum for most plants. Despite the reputation as a bleak and barren soil, cactus soil is actually high in macro and mirco nutrients as well as many vital minerals. Cactus Soil is Coarse Soil mixed for cactus plants is a coarse mixture. Where ordinary soil may have as much as a 20 percent sand content, cactus soil is quite a bit higher, with a sand content of 33 percent or more. For turnips or onions, this type of soil means that the tubers and bulbs can grow larger without fighting through dense soil. The benefits of coarse soil are numerous, including giving life forms in the soil ample ability to move around. Earthworms and insects are vital to soil health, but they often having trouble moving in clay or some other hard-packed soil. Cactus Soil is Rich Seeds will germinate well in moist cactus soil. Because it is so high in vital nutrients and minerals, seeds and seedlings are able to get the energy they need for strong growth. After germination, the seeds can be transplanted to a more loamy soil or they can be left in the cactus soil indefinitely. This soil has everything needed for full grown plant and starting seed alike. It is not as rich as peat alone, however this type of soil contains more nutrients than are available from clay or sandy soils. Mixing Lime into Cactus Soil Sometimes, the acidity level of cactus soil may be too high for less tolerant plants to handle. When this happens, the solution is to add lime, a natural acid neutralizer. You don't need to use a lot of lime and most experts recommend that lime should only be applied to cactus soil in increments of 1 part for every 10 parts of soil. Many clay soils can neutralize soil acidity as well and many experts prefer the use of Gypsum to lime. Cactus Soil Drains Well Because of the high sand content, cactus soil is recognized for good drainage. For garden plants that do not grow well in a lot of water, especially bulbs and tubers, cactus soil is a popular choice. It will retain moisture, but the soil will drain before water levels are able to build up around the roots of the plants. If if you do not use cactus soil for the plant itself, having it as the bottom layer in a planter or flower pot will help you regulate the moisture in the container. Cactus Soil for Planters In planters, cactus soil is excellent for most plants. Whether you intend to grow shrubs or herbs, this type of soil is dense enough to support a large plant, yet loose enough to allow roots to grow unimpeded. It contains a beneficial mix of Nitrogen, Phosphorus, and Potassium and includes most of the micro nutrients required by plants as well.
2.541432
m-a-p/FineFineWeb
Participants consider the water features they might enjoy at a community park — a pond, brook, water playground (or "sprayground"), or pool, — and what happens to thewater over time. 0 Check It Out How-to Video Implementation Guide Provides extensive background information, facilitation outline, materials shopping list, extended supporting media suggestions, correlations to national standards, and more. Teacher's Guide Provides classroom connections, key concepts, connections to science standards, and additional resources.
2.350668
openbmb/Ultra-FineWeb
Enrum is a listed mansion on Strandvejen in Vedbæk, Rudersdal Municipality, some 20 kilometres north of Copenhagen, Denmark. The history of the property dates back to 1731 but the current main building was built in the 1860s to a design by Johan Daniel Herholdt. Enrum is today owned by nemlig.com founder Stefan Plenge. In 1731, royal cabinet secretary Georg Christian Jacobi acquired a piece of land from princess Sophia Hedwig on the coast north of Copenhagen and constructed a small country retreat at the site for his own use. The name Enrum is first recorded in 1733. In 1745, Jacobi sold the house to Michael Fabritius, a co-founder of the Danish Asia Company as well as Kurantbanken. After his death in 1845, his wife, Anna Maria de Longueville, increased the size of the estate through acquisition of new land. After her death in 1775, their son, Conrad Fabritius, bought out the other heirs. He built a new main building and spend considerable sums on the park. In 1778, he was ennobled under the name Fabritius de Tengnagel. In 1805, Fabricius' widow sold Enrum to Christian Colbjørnsen, Denmark's Chief Justice since 1902. In 1811 he sold Enrum to Peder Andreas Kolderup Rosenvinge, director general of the Royal Danish Postal Services, who only owned the property until 1814. The new owner was Johan Frederik Gyldenstierne Sehested, a military officer, and after his death in 1833 the estate came under Det Sehestedske Fideikommis. a family foundation. Enrum was in 1845 acquired by Count Christian Conrad Sophus Danneskiold-Samsøe, who also owned the Holmegård Estate at Næstved. He commissioned Johan Daniel Herholdt to design a new house which was completed in 1864. Count Danneskiold-Samsøe spent his summers on the estate where the Danish Royal Family and the Russian tsar family were among his guests. Enrum remained in the hands of the Danneskiold-Samsøe family until 1939 when it was sold to the Independent Order of Odd Fellows. They used it as a holiday resort for their members until 1984 when it was sold once again. Enrum was later converted into offices by TK Development. From 2005 it served as company headquarters for Stones Invest. In 2013 it was sold in auction for DKK 72 million. In 2017, Enrum was purchased by nemlig.com-founder Stefan Plenge for DKK 49 million. Enrum Forest has an area of 31 hectares. It adjoins Trørød Forest and is open to the public since it has been leased by the Danish Nature Agency on a 50 years contract. It southern part is covered by a small lake, Enrum Pond (Enrum Dam), a boggy area and meadows. The rest is mainly beech forest. In the western part of the forest, in the central axis of the main building, stands the Friendship Column. In the northern part of the forest stands King Charles' Spring (Kong Carls Kilde), which supposedly catered to King Charles XII of Sweden during the departure of his troops from Vedbæk Harbour in 1700. Enrum features in season 3 of the television series The Killing as the stately home of the Zeuthen Family. List of owners (listen er ikke komplet) - (_PHONE_) Georg Christian Jacobi - (_PHONE_) Michael Fabritius - (_PHONE_) Anna Maria Köster - (_PHONE_) Conrad Fabritius de Tengnagel - (_PHONE_) Debora Kloppenburg - (_PHONE_) Peder Andreas Kolderup Rosenvinge - (_PHONE_) Christian Colbjørnsen - (_PHONE_) Johan Frederik Gyldenstierne Sehested - (_PHONE_) Det Sehedstedske Fideicommis - (_PHONE_) Christian Conrad Sophus Danneskiold-Samsøe - (_PHONE_) Ernest Danneskiold-Samsøe - (_PHONE_) Aage Conrad Danneskiold-Samsøe - (_PHONE_) Odd Fellow Ordenen - (_PHONE_) Ukendte ejere - (2005-2007) TK Development - (2007-2012) Steen Gude, Stones Invest - (2012-2013) Husmer Holding - (2013-2014) Slottet Enrum A/S - (2014-2016) Finansiel Stabilitet - (2016 -) Signe Rosenvinge Thürmer & Stefan Plenge - Danish Wiki nemlig.com - "På vandretur - Vedbæks landsteder" (PDF) (in Danish). vores-fam.dk. Retrieved 2014-01-31. - "Enrum" (in Danish). Gyldendal. Retrieved 2014-01-31. - "Conrad Fabritius de Tengnagel" (in Danish). Gyldendal. Retrieved 2014-01-31. - "Bedrageridømte Steen Gudes slot solgt på tvangsauktion for 72 millioner kr" (in Danish). Estate Media. Archived from the original on 2014-02-01. Retrieved 2014-01-31. - "Sag: Enrum" (in Danish). Kulturstyrelsen. Retrieved 2014-01-31. - "Enrum Skov" (in Danish). Rudersdal Municipality. Archived from the original on 2014-02-02. Retrieved 2014-01-31. - "Enrum" (in Danish). Naturstyrelsen. Archived from the original on 2014-02-10. Retrieved 2014-01-31. - "Enrum Slot i Vedbæk lægger kulisse til DR's successerie "Forbrydelsen"" (in Danish). Vedbæk Nyt. Retrieved 2014-01-31. |Wikimedia Commons has media related to Enrum.|
1.289802
m-a-p/FineFineWeb
Vegetable gardens need to be fertilized before planting to keep the soil from being sapped of nutrients. A homemade plant food allows the plants in your garden to produce strong, healthy vegetables. Simply spread a layer of it on top of the prepared garden bed, along with a thin layer of manure, and mix it into the soil with a hand tiller. The standard measurement for this type of plant food is 4 to 6 qts. of fertilizer mix for every 100 square feet of garden plot. All of the ingredients are available at your local gardening center. Wash and dry a large container with a lid. You'll use this container both for mixing and storing the homemade vegetable food. Pour 4 parts seed meal, 1/4 part agricultural lime, 1/2 part dolomitic lime and 1/4 part gypsum into the container. Add 1 part kelp meal and 1 part bone meal to the mixture. Mix the homemade vegetable garden food thoroughly, evenly dispersing the ingredients throughout the mixture. Snap the lid onto the container and store the plant food in a cool, dry spot until it's time to apply it to your garden plot.
1.274581
m-a-p/FineFineWeb
between research scale use and consumer technology scale adaptation of this technology. In this thesis we will explore many, and overcome a few of these challenges. We will address the poor conductivity at the narrow nanowire-nanowire junction points in the network by developing a laser based process to weld nanowires together on a microscopic scale. We address the need for a comparative metric for transparent conductors in general, by taking a device level rather than a component level view of these layers. We also address the mechanical, physical, and thermal limitations to the silver nanowire networks by making composites from materials including a colorless polyimide and titania sol-gel. Additionally, we verify our findings by integrating these processes into devices. Studying a hybrid organic/inorganic heterojunction photovoltaic device we show the benefits of a laser processed electrode. Green phosphorescent organic light emitting diodes fabricated on a solution phase processed silver nanowire based electrode show favorable device metrics compared to a conductive oxide electrode based control. The work in this thesis is intended to push the adoption of silver nanowire networks to further allow new device architectures, and thereby new device applications. 4. Review on measurement techniques of transport properties of nanowires. PubMed Rojo, Miguel Muñoz; Calero, Olga Caballero; Lopeandia, A F; Rodriguez-Viejo, J; Martín-Gonzalez, Marisol 2013-12-01 Physical properties at the nanoscale are novel and different from those in bulk materials. Over the last few decades, there has been an ever growing interest in the fabrication of nanowire structures for a wide variety of applications including energy generation purposes. Nevertheless, the study of their transport properties, such as thermal conductivity, electrical conductivity or Seebeck coefficient, remains an experimental challenge. For instance, in the particular case of nanostructured thermoelectrics, theoretical calculations have shown that nanowires offer a promising way of enhancing the hitherto low efficiency of these materials in the conversion of temperature differences into electricity. Therefore, within the thermoelectrical community there has been a great experimental effort in the measurement of these quantities in actual nanowires. The measurements of these properties at the nanoscale are also of interest in fields other than energy, such as electrical components for microchips, field effect transistors, sensors, and other low scale devices. For all these applications, knowing the transport properties is mandatory. This review deals with the latest techniques developed to perform the measurement of these transport properties in nanowires. A thorough overview of the most important and modern techniques used for the characterization of different kinds of nanowires will be shown. PMID:24113712 5. Thermal and transport properties of a single nickel nanowire NASA Astrophysics Data System (ADS) Ou, Min-Nan; Chen, Yang-Yuan; Wu, Maw-Kuen; Yang, Tzong-Jer; Lee, P. C.; Harutyunyan, S. R.; Chen, C. D.; Lai, S. J. 2008-03-01 Starting with a 100 nm nickel film grown on a Si3N4/Si substrate by thermal evaporator, a suspended nickel nanowire (Ni-NW) was fabricated through e-beam lithography and etching processes. The Ni-NW was a part of 4-probes circuit which is designed for electrical, thermal and thermopower measurements. The resistivity (ρ) and thermal conductivity (κ) of a single nickel nanowire have been measured in the temperature range from 4 to 300 K by 4-probes method and the self-heating-3φ technique. At 300 K the thermal conductivity of nanowire is ˜ 20% of the bulk, it diminishes to lower value as temperature decreases, the consequence is opposite to that in the bulk in which it decreases with temperature increase. The result might be explained by the restriction of mean free paths of electron/phonon--phonon interactions due to the grain boundaries. The small relative resistivity ratio (RRR ˜ 2) confirms the polycrystalline characteristic of the nanowire. The thermopower (Seedbeck coefficient S) was also investigated by temperature gradient built up between two ends of the nanowire. The figure of merit ZT=S^2σ/κ in the one-dimension specimen will be discussed. 6. Review on measurement techniques of transport properties of nanowires. PubMed Rojo, Miguel Muñoz; Calero, Olga Caballero; Lopeandia, A F; Rodriguez-Viejo, J; Martín-Gonzalez, Marisol 2013-12-01 Physical properties at the nanoscale are novel and different from those in bulk materials. Over the last few decades, there has been an ever growing interest in the fabrication of nanowire structures for a wide variety of applications including energy generation purposes. Nevertheless, the study of their transport properties, such as thermal conductivity, electrical conductivity or Seebeck coefficient, remains an experimental challenge. For instance, in the particular case of nanostructured thermoelectrics, theoretical calculations have shown that nanowires offer a promising way of enhancing the hitherto low efficiency of these materials in the conversion of temperature differences into electricity. Therefore, within the thermoelectrical community there has been a great experimental effort in the measurement of these quantities in actual nanowires. The measurements of these properties at the nanoscale are also of interest in fields other than energy, such as electrical components for microchips, field effect transistors, sensors, and other low scale devices. For all these applications, knowing the transport properties is mandatory. This review deals with the latest techniques developed to perform the measurement of these transport properties in nanowires. A thorough overview of the most important and modern techniques used for the characterization of different kinds of nanowires will be shown. 7. Synthesis of Ultrathin Copper Nanowires Using Tris(trimethylsilyl)silane for High-Performance and Low-Haze Transparent Conductors. PubMed Cui, Fan; Yu, Yi; Dou, Letian; Sun, Jianwei; Yang, Qin; Schildknecht, Christian; Schierle-Arndt, Kerstin; Yang, Peidong 2015-11-11 Colloidal metal nanowire based transparent conductors are excellent candidates to replace indium-tin-oxide (ITO) owing to their outstanding balance between transparency and conductivity, flexibility, and solution-processability. Copper stands out as a promising material candidate due to its high intrinsic conductivity and earth abundance. Here, we report a new synthetic approach, using tris(trimethylsilyl)silane as a mild reducing reagent, for synthesizing high-quality, ultrathin, and monodispersed copper nanowires, with an average diameter of 17.5 nm and a mean length of 17 μm. A study of the growth mechanism using high-resolution transmission electron microscopy reveals that the copper nanowires adopt a five-fold twinned structure and evolve from decahedral nanoseeds. Fabricated transparent conducting films exhibit excellent transparency and conductivity. An additional advantage of our nanowire transparent conductors is highlighted through reduced optical haze factors (forward light scattering) due to the small nanowire diameter. 8. Indium tin oxide nanowires grown by one-step thermal evaporation-deposition process at low temperature. PubMed Dong, Haibo; Zhang, Xiaoxian; Niu, Zhiqiang; Zhao, Duan; Li, Jinzhu; Cai, Le; Zhou, Weiya; Xie, Sishen 2013-02-01 Indium tin oxide (ITO), as one of the most important transparent conducting oxide, is widely used in electro-optical field. We have developed a simple one-step method to synthesize ITO nanowires at low temperature of 600 degrees C. In detail, mixtures of InN nanowires and SnO powder, with the molar ratio of 10:1, have been used as precursors for the thermal evaporation-deposition of ITO nanowires on silicon/quartz slices. During the growth process, the evaporation temperature is maintained at 600 degrees C, which favors the decomposition of InN and oxidation of In, with a limited incorporation of Sn in the resulting compound (In:Sn approximately 11:1 in atomic ratio). As far as we know, this is the lowest growth temperature reported on the thermal deposition of ITO nanowires. The diameters of the nanowires are about 120 nm and the lengths are up to tens of micrometers. XRD characterization indicates the high crystallization of the nanowires. HRTEM results show the nanowires grow along the [200] direction. The transmittance of the nanowire film on quartz slice is more than 75% in the visible region. Based on photolithography and lift-off techniques, four-terminal measurement was utilized to test the resistivity of individual nanowire (6.11 x 10(-4) omega x cm). The high crystallization quality, good transmittance and low resistivity make as-grown ITO nanowires a promising candidate as transparent electrodes of nanoscale devices. PMID:23646624 9. A comprehensive study of thermoelectric and transport properties of β-silicon carbide nanowires SciTech Connect Valentín, L. A.; Betancourt, J.; Fonseca, L. F.; Pettes, M. T.; Shi, L.;
1.418184
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
- PASSENGER-STRAND (1) (remove) - A large-scale chemical modification screen identifies design rules to generate siRNAs with high activity, high stability and low toxicity (2009) - The use of chemically synthesized short interfering RNAs (siRNAs) is currently the method of choice to manipulate gene expression in mammalian cell culture, yet improvements of siRNA design is expectably required for successful application in vivo. Several studies have aimed at improving siRNA performance through the introduction of chemical modifications but a direct comparison of these results is difficult. We have directly compared the effect of 21 types of chemical modifications on siRNA activity and toxicity in a total of 2160 siRNA duplexes. We demonstrate that siRNA activity is primarily enhanced by favouring the incorporation of the intended antisense strand during RNA-induced silencing complex (RISC) loading by modulation of siRNA thermodynamic asymmetry and engineering of siRNA 3-overhangs. Collectively, our results provide unique insights into the tolerance for chemical modifications and provide a simple guide to successful chemical modification of siRNAs with improved activity, stability and low toxicity.
1.435713
HuggingFaceFW/fineweb-edu
Monday, March 10, 2008 Income is not income Money: A song written by Berry Gordy, first recorded by Barrett Strong, then a bit later by a British band called the Beatles. But income is not income. If you check the results of the Census/ACS and BEA for similar time frames, you'll see two different definitions of income. The Census/ACS definition is MONEY income. BEA measures PERSONAL income. Data Detective Lenny explains: Money income is essentially gross cash received on a regular basis (inheritence, etc. are not included). So wages and salaries are measured before taxes are removed. According to the BEA web site: This includes such income as housing subsidies, Medicare/Medicaid payments to hospitals on behalf of individuals, etc. even though the individual does not personally see it. The idea is that this is money supporting individuals. So the BEA figures will be significantly different ($10k higher) then the Census Bureau's. No comments:
1.15898
Zyphra/Zyda-2
,,,, Nour: Play With Your Food's sous chef: an interactive soundtrack Reading Time: 6 minutes No two dishes made in Nour: Play With Your Food are alike. A delicious interactive exploration into the aesthetics of food and drink, Nour lets you play with your food in ways your mother told you not to. This beautiful, chaotic yet comforting game demanded a soundtrack that reflects each player's unique preparations and plating. What started as a soundtrack ended up as an instrument. Play sweet melodies and manipulate the environment via random acts of cookery, rhythmic prowess, or even your singing voice. Beginner or expert, reckless or meticulous, every chef creates a song to match their pièce de résistance. I'm Maximilian C Mueller, a musician and designer – and today, your maître d'Nour's music. James Morrison (aka Fluke Nukes) and I cooked up a feast for the ears to accompany this feast for the eyes. Ambience: Season to Taste Adaptive background music to complement your dish. The music heats up, cools down, rises, mixes, and melts. What you hear is a unique combination of ingredients cooked per ever-changing specifications set by the way you play; no two play-throughs will sound the same. Wreak havoc, show restraint, and so too will the audio. Adaptive game music is often categorized as "horizontal" (e.g. the music changes when entering a room), or "vertical" (e.g., tempo increases or an instrumental element is added when entering combat). Nour's music is diagonal and has a cherry on top. Player actions and their frequency comprise a measure of chaos used to generate a cohesive song from a large pantry of ingredients. Together with a dash of stochasticity and a pinch of salt, this fluctuating measure acts as a conductor, guiding the scene sound. Put simply, the dish you create creates the song. Have a listen to some evolving music captured in-game while you read: Marshmallow Flutes and Noodle Harps Conjure food and make a melody, create a dish and make a song. Each time a food is introduced into the scene, a sound comes with it. What does a marshmallow sound like? An airy flute! An ice cube? The high notes of a grand piano played staccato of course! It was great fun choosing these sounds and even more fun subjecting those sounds to the trials of gameplay. Don't worry about playing sour notes; every button press is always in tune. Once in the scene, the foods act as percussion instruments, singing the sound that announced them on impact. I love setting up a pile of ice cubes and slamming something into them to create an icy physics music explosion. Perhaps the best example of physical instrumentation in Nour is the noodle harp. Make some noodles, arrange them to form an elegant slope (or monster) using the magnet, and freeze them. While there are many ways you could choose to play the harp, one of my favorites is simply letting a gyoza tumble down the sculpture to create a descending melody. Making a beat in Toaster using a combination of "broken" and working toasters. Each scene in Nour has some unique gameplay mechanics and musical mechanics to match. In Toaster, make toast and conduct an orchestra at the same time. Bread takes to the air upon a swell of violins, bassoons, and timpani. Burn your toast to the degree that the toaster breaks…and what was once a bassoon is a BASS. Appliance destruction triggers a change of temperament in your musicians, turning the smoldering toasters into 808's, snares, hats, and kicks. Create a fiery beat – but be sure to leave one or two toasters unbroken for a nice orchestral accompaniment. Fire dances to the beat when night falls in Toaster. In Meat Grinder, each food is spawned into a calm but foreboding environment. Things really take a turn when the player rotates the handle to grind a food – not only is the food ground, but its sound is extruded too. There may or may not be heavy-metal involved. Players can look forward to both macro and micro manipulation of game audio via changes they make to the environment. This relationship is reciprocal: they'll also have the power to change the environment through music. A jazz cover of the Miso Soup music hidden in an easter egg – happy hunting! Sonic Controller: Audio-as-Interface Use music to manipulate food. Music in Nour acts not just as a soundtrack, but an interface. Players can exert influence over the scene and its contents in both familiar and new modes of musical gameplay. The mechanics aim to encourage and reward the exploration and experimentation central to Nour. Fittingly, its rhythm system has no grid. User input is instead processed relatively, allowing the game to recognize and reward melodies ranging from rigid to adventurous. Rhythmic play generates a resource that can be used to activate special powers: freeze foods to create a sculpture (or plate), turn day into night, or compel your food to dance to the music. Combining these effects creates a dizzying number of musical (and physical) possibilities. Lock some dancing sugar cubes in a cage of frozen oranges to create a dynamic physical instrument – the sugar pulses to the music, knocking into itself and the walls confining it. Then activate night mode to really get things bopping! I'm so excited to see what players come up with and hear the sounds they're able to coax out of the scenes. Secret rhythmic sequences (combos) offer another way to control the game – I'll give away one to get you started: try slowing down time by playing 4 quarter notes, 2 halves, and a whole. And for balance, an anti-sequence: button mash as fast as your can to activate Warp Mode! We hope such rhythmic mechanics will be fun, familiar, and legible to those who have stomped, shredded, or ocarina-ed their way through a virtual space. The DualSense controller, Sequence effects are triggered by specific note sequences. This sequence, (4, 4, 4, 4, 2, 2, 1), prompts a key change and slows down both the music and gameplay. however, furnished an opportunity for something new. Not only can you play with your food, you can yell, sing, blow, and slurp at it. Sing the root note of the song to levitate objects upon a field of golden energy for as long as you can hold it. Blast away objects with a gust of wind blown into the microphone. Yell to charge up a food and plate it with force! After all that you'll likely need to take a sip of refreshing soda – just produce a rising pitch with your voice, a cello, or…? Left: Using voice to levitate citrus in Soft Drinks // Right: Blowing a gust of wind in Diner Panic's previous title, Untitled Goose Game, was praised for its reactive score, so from the start we knew Nour would have to push the audio envelope. This is our first time working on a game, so we didn't really know what was possible in the key of C# – it turns out a lot is! It's been a wild ride up a steep learning curve – a big thanks to our very talented, supportive, and patient team for sticking with us while we skilled up. We're all so excited for you to experience Nour: Play With Your Food this summer. Website: LINK Facebook Comments Surreal narrative adventure Where the Heart Leads hits PS4, PS5 on July 13
1.292418
Zyphra/Zyda-2
EN CASO DE ERROR 28 Descargar (0) Texto completo (1) PL/SQL - En caso de error • Bloques - Expresiones • Registros • Tablas • Tablas - Conversiones de datos • Variables y constantes • Cursores - Excepciones - Estructuras de control EN CASO DE ERROR Q é h d ? Š ¿Qué hacer en caso de error? „ "La base de datos no funciona bien, no hace lo que quiero" „ Proceso de resolución de errores: 1. Revisar la tabla user_errors 1. Show errors 2. Select * from user_errors; Si i ? 2. ¿Sintaxis correcta? 3. ¿Problema con una consulta? 1. La consulta no devuelve resultados 2. Fuera del disparador la consulta funciona pero dentro no (2) INTRODUCCIÓN (I) PL/SQL L j d ió di l Š PL/SQL:Lenguaje de programación procedimental estructurado en bloques que amplía el lenguaje estándar SQL. Š Permite: „ Manipular datos de una BD Oracle. „ Usar técnicas procedurales (bucles, ...) „ Controlar las filas de una consulta una a una Controlar errores (excepciones) definidas por el usuario o propios de - 3 Fundamentos de las Bases de Datos „ Controlar errores (excepciones) definidas por el usuario o propios de Oracle (predefinidos) „ Disparadores Š No diferencia las minúsculas de las mayúsculas CLIENTE == cliente BLOQUES Š Es la unidad básica de cualquier programa PL/SQL. Š Estructura básica de un bloque (sólo es obligatorio el conjunto de sentencias ejecutables): DECLARE /* Declaraciones de uso local: variables, cursores, y excepciones de usuario */ BEGIN /* Proceso: conjunto de sentencias ejecutables */ EXCEPTION /*Excepciones:zona de control de errores */ (3) BLOQUES (II) Š Las únicas instrucciones permitidas dentro de un p bloque son: INSERT, UPDATE, DELETE Y SELECT + manipulación de datos + control de transacciones. Š La anidación de bloques se realiza entre las etiquetas BEGIN Y EXCEPTION y sólo se tiene una definición de variables. I t i itid DROP - 5 Fundamentos de las Bases de Datos Š Instrucciones no permitidas son: DROP, CREATE, ALTER, … Š Los comentarios se ponen: /* Comentarios */ Š No distingue mayúsculas y minúsculas BLOQUES (III) Š Tipos: „ Anónimo: construido de manera dinámica y se ejecuta una vez. z Estructura básica se compilan cada vez que son ejecutados y no se guardan en la BD. „ Nominado:bloque anónimo con etiqueta. Lo primero es la etiqueta. PE -> <<insertarEnTabla>> z Subprogramas:procedimientos, paquetes y funciones almacenadas en la BD ya compiladas. Se ejecutan múltiples veces mediante llamadas explícitas. z Remplazar la palabra clave DECLAREcon las palabras CREATE OR REPLACE PROCEDURE nombre_proc AS „ Disparadores:bloques nominados almacenados en la BD (código (4) EXPRESIONES Ti d i Š Tipos de expresiones: „ Aritméticas : + - * / „ Comparaciones : = != > > >= <= „ Concatenación de caracteres ⏐ ⏐ Š Tipos de variables: „ Escalares. Definidos por el lenguaje NUMBER CHAR VARCHAR VARCHAR2 DATE - 7 Fundamentos de las Bases de Datos z NUMBER, CHAR, VARCHAR, VARCHAR2, DATE, BOOLEAN (TRUE, FALSE, NULL) „ Compuestos: Definidos por el usuario z Registros z Tablas y matrices. Pueden almacenar registros y escalares REGISTROS R i D fi i ió Š Registros. Definición TYPE tipo_reg IS RECORD (campo1tipo1[NOTNULL][:=expr ];... campontipon[NOTNULL][:=expr ];); Š Registro. Creación de variables Nombre_variable tipo_reg Š Referenciar las variables Nombre_variable.campo:=valor Š Ej l Š Ejemplo DECLARE (5) TABLAS T bl d fi i ió Š Tablas definición. „ TYPE tipo_tabla IS TABLE OF tipo INDEX BY BINARY_INTEGER; Š Tablas. Definición de variables: vcliente tipo_tabla; Š Tablas. Referencia: Tabla(indice).campo; Š Tablas. Ejemplo DECLARE TYPE ttabla IS TABLE OF cliente%ROWTYPE - 9 Fundamentos de las Bases de Datos TYPE ttabla IS TABLE OF cliente%ROWTYPE INDEX BY BINARY_INTEGER; v_cliente ttable; BEGIN SELECT * INTO v_cliente(_PHONE_) FROM cliente WHERE DNI=22522480; DBMS_OUTPUT.PUT_LINE (vcliente(_PHONE_).nombre); END; Tablas. Atributos Š Tablas. Atributos Atributo Tipo Devuelto Descripción COUNT NUMBER Devuelve el número de filas de la tabla DELETE (nº) N/A Borra las filas de una tabla EXITS (nº) BOOLEAN Devuelve TRUE si existe en la tabla la fila especificada FIRST BYNARY_ INTEGER Devuelve el índice de la primera fila LAST BYNARY_ INTEGER Devuelve el índice de la última fila NEXT BYNARY_ INTEGER Devuelve el índice de la fila de la tabla que sigue a la fila especificada PRIOR BYNARY_ INTEGER (6) CONVERSIONES DE DATOS Š Entre escalares se pueden convertir datos de distintas p familias excepto las propias restricciones de las variables (Ej. CHAR (10) no puede en VARCAHAR2(1)) Función Descripción TO_CHAR Convierte el argumento en tipo VARCHAR2 - 11 Fundamentos de las Bases de Datos TO_DATE Convierte su argumento en tipo DATE TO_NUMBER Convierte el argumento en tipo NUMBER Declaración de variables y constantes Š Variables: Se utilizan para almacenar valores devueltos por una consulta o para realizar cálculos intermedios consulta o para realizar cálculos intermedios. Š Constantes: Son campos definidos e inalterables Š Pasos a seguir z Definición Nombre_campo: nombre de la variable z Declaración Nombre_campo tipo [CONSTANT][NOTNULL] [:=VALOR]; z Asignación Variable Tipo. Un identificador es el nombre de la vble. z Tipo: puede ser: Š Tipo de datos: tipo de dato de la variable Š Identificador%TYPE Se refiere al tipo usado por una columna) Š Identificador%ROWTYPE Es una fila vble con los mismos nombres y tipos que las y p q columnaas de una tabla o fila recuperada de un cursor) Š [CONSTANT] Palabra reservada para def. de ctes Š [NOTNULL]: Obliga a tener valor (7) CTES Y VBLES. EJEMPLO DECLARE DECLARE DNI NUMBER (8,0); Nombre VARCHAR (30); Factor CONSTANT NUMBER(3,2):=0.10; DNI2 cliente.DNI%TYPE; Rcliente cliente%ROWTYPE; (tendría los campos: Rcliente DNI Rcliente Nombre ) - 13 Fundamentos de las Bases de Datos (tendría los campos: Rcliente.DNI, Rcliente.Nombre …) precio NUMBER:= 300; (inicializa a un valor) CURSORES Š Definición: „ Cursor es un área de trabajo definida para las consultas que devuelven más de una fila. „ Permite la manipulación de datos „ Se le pueden pasar parámetros Š Tipos de
1.0208
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
What does the upstream for major packages really look like? Over the past few years, the shape of the open source ecosystem landscape has shifted drastically, exploding both in the volume of published code, and also the number of dependencies that live upstream from a given library. Scoping the Discussion In order to have a real discussion about software packages and ecosystem health, it's important to first understand the current state of the ecosystem. For the sake of simplicity, we'll keep the focus of this discussion to NPM - which represents the logical extreme of the problems and ideas discussed here. Keep in mind, however, that this is absolutely not exclusively a Javascript issue - a cursory glance shows that all of these problems are just as present and relevant in every other programming language ecosystem - even if the awareness is somewhat lower. Some Background In the 90's, the major trends surrounding software development all centered around one core concept: Code Reuse. While code reuse has probably been a part of software development from the earliest days of industry, the practices around this period began to favor it very heavily - during this decade, languages and APIs built around Object Oriented (OO) concepts started to really rise to prominence, and the earliest iterations of package managers were minted. Interface definitions and library packaging were at the forefront of the tooling. If we fast forward to the end of the 90's, specifically to the year 1999 - the year when The Pragmatic Programmer was published - we see this whole concept codified with the DRY (Don't Repeat Yourself) principle. This, to be entirely fair, is a great thing for software - reusing previously-written code (that has already been tested, validated, and vetted) probably saves billions in development costs annually. On the other hand, however, it does come with some implicit risks. Software Reuse: The Implications While all of this was not only very well intentioned, but also a generally smart way to think about the problem, the landscape has shifted drastically since then. Gone are the days when the people behind the software you consume as part of your build process are really understandable - the scope and scale of the open source ecosystem have ballooned over the last few years, and continue to do so at every turn. The implication here is two-fold: Firstly, it is difficult to really reason about what you are incorporating into your codebase when leveraging an external library, giving rise to a whole host of paid solutions that simply provide you with a "bill of materials," and second, there are a whole host of hidden costs bundled into the upstream packages and libraries. The technical debt contained in those packages becomes your technical debt, their issues (to varying degrees) become your issues. This also leads to a great deal of what-ifs: what if one of those packages has a critical vulnerability? What if one of them contains licenses that may impact your commercial viability? What if some of them are abandoned? Contain malware? Have backdoors? The short answer is: All of these are valid concerns. Some of them have come to the forefront of the public focus (particularly issues around licenses and vulnerabilities), but as for the rest - they represent a veritable ticking time bomb in the community; one which has arguably gone off already in the form of 700 malicious RubyGems, hundreds to thousands of de-listed npm packages, and dozens of Python libraries, and tens of Docker images, among others (while I did promise up front that we'd be focused on NPM, it very much bears mentioning the scope of these issues in other, tangential ecosystems). Rough Numbers So even if we believe that all of this is a big issue, that begs the question: How big is it? Just to get a rough idea, this "State of the Union" post about NPM (published in early 2017) indicates that in 2015 - 5 years ago (at the time of writing) - NPM contained a scant 12,500 packages. Fast forward to today, and it contains nearly 1.4 million, and continues to grow at an average rate of almost 900 a day. npm growth rate NPM Growth Trajectory - Graph courtesy of modulecounts.com Keep in mind that this is net new packages - not at all counting updates to existing ones under active development. What does this really mean, though? Do we really care about the size of the ecosystem if we only use one or two major packages during development? In order to get a better understanding here, we did some more detailed analysis on React (an immensely popular UI framework, for those unfamiliar) to figure out what its dependency graph looks like. A first pass at this seemed like a good spot to simply look at some products that do commodity "bill of materials" and vulnerability analysis. Picking a popular tool, we checked out the latest react build (at the time of writing), and looked to see what its dependency tree would look like: React Dependencies Misleading Proported React Dependency Count At first glance, this doesn't really seem to tell the whole story; if we pull open react's package.json directly, we also see this: react dev dependencies ... And on for tens more lines The "devDependencies" are totally ignored! While this may not seem like a big deal at first blush (they aren't included in the final build, after all!), in many ways, they actually pose an even bigger security threat than the production dependencies. Production dependencies for React run in the web browser's sandbox - and while there is certainly a whole host of bad stuff that can be done there, dev dependencies have the ability to interact directly with the host they are running on without any sandbox, meaning they are free to access files, connect to remote servers, and run additional programs with the same permissions as the current user. Really, at its core this means that fundamentally untrusted code, written by random developers, runs on sensitive systems where security credentials are stored, such as developer workstations, CI runners, and more. If stealing some customer data here and there from the browser seems really bad, what does the ability to steal any keys, credentials, and access that a developer or CI runner has the ability to access mean? Most developers have root access to their own system, and often have access to sensitive infrastructure and data, including production systems. CI runners are often worse - modern CI/CD processes frequently inject credentials to everything during the build and deployment process, giving an attacker the proverbial "keys to the kingdom". To make matters even worse, simply looking at the package.json doesn't really tell the whole story - nearly all of the dependencies in that list in turn have dependencies, which in turn have more dependencies, and so on. If we pull down the entire graph just for React, we get something like 7,000 total dependencies for a single library. react full dependency graph The real dependency graph To be totally clear, this is not what you'll get every time you run npm install, but rather a superset of what you'll get during that process. Due to semantic versioning (notionally, the option to specify ranges of acceptable package versions in your package.json which will satisfy a dependency), it is all but impossible to say exactly what you'll get a priori, as large parts of the dependency graph may change substantially between package versions. Really, all of this simply compounds the root problem: well meaning development principles have been pushed to their logical extreme. This ultimately makes simply understanding what you'll get with a given install an almost impossibly hard problem. This is doubly true if you want to develop anything more than a cursory understanding of the stuff that sits upstream from your development efforts - not mentioning the authors and maintainers behind those packages. Not Just Security While security concerns are certainly at the forefront of this discussion, a whole separate class of issues presents itself here: How does the code quality of upstream packages impact the technical debt of your projects? While this seems like sort of a subjective thing to measure on its face, some clear problems present themselves: Does the repository in question have good test coverage? Does it have unit tests at all? What about the contributors, how responsive are they to issues? How many contributors have the ability to commit code - if it is only one, what will happen if that contributor stops working on the project? And as a follow on, does the repository appear to be abandoned? These are certainly all concerns for every consumer of open source - issues in any of these areas in a single library will be issues for every package that depends upon them. It Only Gets Worse While all of this seems bad enough, what about updates to those 7,000 existing packages? Even if we have undeniable proof that all of the authors sitting upstream are honest and well intentioned, what happens if they suffer a credential breach, like the relatively-recent one suffered by Docker Hub? Additionally, what if their account credentials are compromised in some other way? Either scenario puts every package downstream completely at the mercy of whoever obtains their credentials. This may seem like something of a hypothetical situation, but that is unfortunately not the case; doing some very high-level sampling of some top contributors (with hundreds of published packages, all upstream from dozens of major projects) and comparing the results with some major password breach datasets yielded disturbing results: the very first such contributor we looked at appeared in no less than three such datasets, with the same password in use in all cases - a password extremely vulnerable to dictionary attacks, no less. At this point, not only do we need to be able to keep track of net new additions to our dependency graph, but also any changes to the thousands to tens of thousands of libraries that exist upstream, as such changes may introduce malware (if the author
1.472983
Zyphra/Zyda-2
A science blog with Eric Berger Bacteria turn deadlier in space. What about babies? The big news today is that Salmonella bacteria shot into space last summer came home three times as deadly. Scientists had expected this because ground-based experiments suggested the zero-gravity environment of space might increase virulence. Would a space-born baby survive? But it's fascinating all the same. All life that we know of evolved in one Earth gravity. We've already seen what effect spaceflight has on human bones, muscles and the immune system (it weakens all of them). Now we're seeing that it has a striking effect even on the smallest of living things. Perhaps the coolest part of the new research is that it pinpoints the changes in Salmonella down to the genetic level, finding that microgravity caused the bacteria to express (turn on/off) 167 different genes than bacteria that were kept on the ground. These genetic changes increased the virulence of the bacteria. While reporting the story, Jeanne Becker at the National Space Biomedical Research Institute told me about an earlier space experiment during which astronauts grew human kidney cells on orbit. Those cells expressed a remarkable 1,600 different genes than the Earth-grown variety. All of this leads me to an intriguing question. If genes express differently in space, what will happen if (when, really) a baby is conceived in space? How about a newborn's development? I can see this becoming a huge moral question some day, especially with the radically different gene expression. A space baby would very likely be something that's virtually non-human, if he or she survived at all. Categories: Genetics, Space Eric Berger 51 Responses 1. ttyler5 says: " …a really expensive way to have fresh omellets…." Oh, Hunter-san, humble servant not think so! Device fix eggs with chemical at diverse stages of embryonic development! Not so tasty and probabrary lethal! " …results? " Firstly, here's a more detailed description of the the device: " Experiment Operations " The Avian Development Facility (ADF) was designed for Japanese quail eggs and houses 36 egg holders. The egg holders have been designed to isolate the eggs from vibration, thus minimizing any launch and re-entry effects to the developing embryos. " Egg holders are mounted on two rotating centrifuges that can provide both microgravity and variable gravity environments (up to 1-g – normal Earth gravity). " The avian facility offers pre-programmable control of the interior environmental temperature, humidity, carbon dioxide (CO2), and oxygen (O2) concentration thus providing optimal conditions for embryo development. " Additionally, it has an automated system that can be programmed to add a chemical fixative to the egg,which stops the development process and preserves the embryo, at specific times during incubation. " The ADF also rotates eggs – similar to the turning in a natural environment. Temperature can be maintained within +/-.05 degree Celsius of the set point between the range of 78.8 degrees Fahrenheit (26 degrees Celsius) and 104 degrees Fahrenheit (40 degrees Celsius). The relative humidity can be maintained within +/- 7.5% of the set point between the range 50 and 70 percent. " The Oxygen and Carbon Dioxide levels are actively controlled to the set points defined by the researcher and the range of centrifugal force provided by the carousels is 0 to 1-g (in 0.1-g increments). " The ADF fits into a middeck locker on the Orbiter, making it easily accessible prior to launch and after landing."
1.81435
Zyphra/Zyda-2
Principal Investigator(s): Summary Tape File (STF) 4C provides summary statistics on population and housing characteristics produced from the 1980 Census. This extract of STF 4C provides data for Standard Metropolitan Statistical Areas (SMSAs), urbanized areas, and the state components of SMSAs and urbanized areas. The summary data are presented as a series of tables, one variable per table cell, which are repeated for each area. Population items tabulated for each area include age, race, sex, marital status, and Spanish origin, plus information on schooling, employment status, children, occupation, and income. Housing items include data on size and condition of the housing unit as well as information on value, age, water, sewage and heating, vehicles, and monthly owner costs. Access Notes Study Description U.S. Dept. of Commerce, Bureau of the Census. CENSUS OF POPULATION AND HOUSING, 1980 [UNITED STATES]: SUMMARY TAPE FILE 4C SMSA EXTRACT. ICPSR08426-v3. Washington, DC: U.S. Dept. of Commerce, Bureau of the Census [producer], 1983. Ann Arbor, MI: Inter-university Consortium for Political and Social Research [distributor], 2006-07-07. doi:10.3886/ICPSR08426.v3 Persistent URL: Export Citation: - EndNote XML (EndNote X4.0.1 or higher) Scope of Study Subject Terms: African Americans, automobile ownership, census data, children, demographic characteristics, education, ethnicity, employment, family, Hispanic origins, households, housing, income, Metropolitan Statistical Areas, occupations, population, property value, states (USA), urban areas Geographic Coverage: United States Date of Collection: - 1980 Universe: All persons and housing units in SMSAs and urbanized areas in the United States. Data Types: aggregate data Data Collection Notes: (1) The data file comprises two record types, Record A and Record B, each with a different series of population and housing tables. Record A is presented for each geographic area and is followed by a number of B records. There is a separate B record for the total population, Spanish origin persons if present in the geographic area, and each race category in the area. The A and B records are divided into segments, each comprising 1,956 characters, which constitute the logical records of the data file. Record A contains 25 segments, 5 with population tables and 20 with housing tables. Record B has 19 segments, 11 with population tables and 8 with housing tables. To protect respondent privacy, the B record for Spanish-origin persons or a race category was omitted if certain suppression criteria were met. (2) In STF 4C, Record B is present for the total population and a maximum of 38 race, Spanish-origin, and ancestry categories. However, this abstract is limited to just seven basic categories: 1. total population, 2. Whites, 3. Blacks, 4. American Indians, Eskimos, and Aleuts, 5. Asians and Pacific Islanders, 6. persons of other races, and 7. Spanish-origin persons. (3) The data file contains 120 to 158 logical records per case and approximately 1,500 variables. The cases are sorted by geographic hierarchy. Sample: Estimates derived from the 1980 Census have been inflated to represent the entire population of the United States. Data Source: 1980 Census of Population and Housing Original ICPSR Release: Version History: - 2006-07-07 Duplicate records (5,218 in number) were removed from the data file and the study description was revised. Metadata Exports Download Statistics
1.804224
Zyphra/Zyda-2
, from a knowledge discovery point of view, the reproduction of known knowledge may not necessarily be the intended result.[32] The F-measure can be used to balance the contribution of false negatives by weighting recall through a parameter \beta \geq 0. Let precision and recall be defined as follows: P = \frac {TP } {TP + FP } R = \frac {TP } {TP + FN} where P is the precision rate and R is the recall rate. We can calculate the F-measure by using the following formula:[30] F_{\beta} = \frac {(\beta^2 + 1)\cdot P \cdot R } {\beta^2 \cdot P + R} Notice that when \beta=0, F_{0}=P. In other words, recall has no impact on the F-measure when \beta=0, and increasing \beta allocates an increasing amount of weight to recall in the final F-measure. J(A,B) = \frac {|A \cap B| } {|A \cup B|} = \frac{TP}{TP + FP + FN} FM = \sqrt{ \frac {TP}{TP+FP} \cdot \frac{TP}{TP+FN} } where TP is the number of true positives, FP is the number of false positives, and FN is the number of false negatives. The FM index is the geometric mean of the precision and recall P and R, while the F-measure is their harmonic mean.[35] Moreover, precision and recall are also known as Wallace's indices B^I and B^{II}.[36] Clustering Axioms[edit] Given that there is a myriad of clustering algorithms and objectives, it is helpful to reason about clustering independently of any particular algorithm, objective function, or generative data model. This can be achieved by defining a clustering function as one that satisfies a set of properties. This is often termed as an Axiomatic System. Functions that satisfy the basic axioms are called clustering functions.[37] Formal Preliminaries[edit] A partitioning function acts on a set S of n \ge 2 points along with an integer k > 0, and pairwise distances among the points in S. The points are not assumed to belong to any specific larger set or space; the pairwise distances are the only data the partitioning function has about them.[citation needed] We may label the points in S using the numbers \{1, 2, \ldots, n\}. The pairwise distances define a distance function d: S \times S \rightarrow R which should have the properties of a semimetric: for any i, j \in S, we must have d(i, j) \ge 0, d(i, j) = d(j, i), and d(i, j) = 0 if and only if i = j. In other words, the distances must be nonnegative, symmetric, and two points have distance zero if and only if they are the same point. A partitioning function F takes a distance function d on S \times S and an integer k \ge 1 and returns a k-partition of S, a collection of non-empty disjoint subsets of S whose union is S. The sets making up the k-partition are the clusters. Two clustering functions are equivalent if and only if they output the same partitioning on all values of d and k. Now in an effort to distinguish clustering functions from partitioning functions, we lay down some properties that one may like a clustering function to satisfy. Here is the first one. If d is a distance function, then we define \alpha \cdot d to be the same function with all distances multiplied by \alpha. For any distance function d, number of clusters k, and scalar \alpha > 0, we have F(d, k) = F(\alpha \cdot d, k) This property simply requires the function to be immune to stretching or shrinking the data points linearly. It effectively disallows clustering functions to be sensitive to changes in units of measurement - which is desirable. We would like clustering functions to not have any predefined hard-coded distance values in their decision process. The next property ensures that the clustering function is "rich" in types of partitioning it could output. For a fixed S and k, Let Range(F(\bullet, k)) be the set of all possible outputs while varying d. For any number of clusters k, Range(F(\bullet, k)) is equal to the set of all k-partitions of S In other words, if we are given a set of points such that all we know about the points are pairwise distances, then for any partitioning \Gamma, there should exist a d such that F(d, k) = \Gamma. By varying distances amongst points, we should be able to obtain all possible k-partitionings. The next property is more subtle. We call a partitioning function "consistent" if it satisfies the following: when we shrink distances between points in the same cluster and expand distances between points in different clusters, we get the same result. Formally, we say that d' is a \Gamma-transformation of d if (a) for all i,j \in S belonging to the same cluster of \Gamma, we have d'(i,j) \le d(i,j); and (b) for all i,j \in S belonging to different clusters of \Gamma, we have d'(i,j) \ge d(i,j). In other words, d' is a transformation of d such that points inside the same cluster are brought closer together and points not inside the same cluster are moved further away from one another. Fix k. Let d be a distance function, and d' be a F(d, k)-transformation of d. Then F(d, k) = F(d', k) In other words, suppose that we run the partitioning function F on d to get back a particular partitioning \Gamma. Now, with respect to \Gamma, if we shrink in-cluster distances or expand between-cluster distances and run F again, we should still get back the same result - namely \Gamma. The partitioning function F is forced to return a fixed number of clusters: k. If this were not the case, then the above three properties could never be satisfied by any function.[38] In many popular clustering algorithms such as k-means, Single-Linkage, and spectral clustering, the number of clusters to be returned is determined beforehand – by the human user or other methods – and passed into the clustering function as a parameter. Biology, computational biology and bioinformatics Plant and animal ecology clustering is used to build groups of genes with related expression patterns (also known as coexpressed genes). Often such groups contain functionally related proteins, such as enzymes for a specific pathway, or genes that are co-regulated. High throughput experiments using expressed sequence tags (ESTs) or DNA microarrays can be a powerful tool for genome annotation, a general aspect of genomics. Sequence analysis High-throughput genotyping platforms clustering algorithms are used to automatically assign genotypes. Human genetic clustering Medical imaging On PET scans, cluster analysis can be used to differentiate between different types of tissue and blood in a three dimensional image. In this application, actual position does not matter, but the voxel intensity is considered as a vector, with a dimension for each image that was taken over time. This technique allows, for example, accurate measurement of the rate a radioactive tracer is delivered to the area of interest, without a separate sampling of arterial blood, an intrusive technique that is most common today. IMRT segmentation Clustering can be used to divide a fluence map into distinct regions for conversion into deliverable fields in MLC-based Radiation Therapy. Business and marketing Market research Grouping of shopping items World wide web Social network analysis Search result grouping Slippy map optimization Computer science Software evolution Image segmentation Evolutionary algorithms Recommender systems Markov chain Monte Carlo methods Social science Crime analysis Educational data mining Field robotics Clustering algorithms are used for robotic situational awareness to track objects and detect outliers in sensor data.[39] Mathematical chemistry To find weather regimes or preferred sea level pressure atmospheric patterns.[41] Petroleum geology Physical geography The clustering of chemical properties in different sample locations. See also[edit] See also[edit] Related methods[edit] 3. ^ Cattell, R. B. (1943). The description of personality: Basic traits resolved into clusters. Journal of Abnormal and Social Psychology, 38, 476-506. 4. ^ a b c d e f Estivill-Castro, Vladimir (June 2002 2002). "Why so many clustering algorithms — A Position Paper". ACM SIGKDD Explorations Newsletter 4 (1): 65–75. doi:10.1145/568574.568575. 5. ^ R. Sibson (1973). "SLINK: an optimally efficient algorithm for the single-link cluster method". The Computer Journal (British Computer Society) 16 (1): 30–34. doi:10.1093/comjnl/16.1.30. 6. ^ D. Defays (1977). "An efficient algorithm for a complete link method". The Computer Journal (British Computer Society) 20 (4): 364–366. doi:10.1093/comjnl/20.4.364. 7. ^ Lloyd, S. (1982). "Least squares
2.385475
Zyphra/Zyda-2
Depth range (m): 0 - 4.8 Depth range (m): 0 - 4.8 Note: this information has not been validated. Check this *note*. Your feedback is most welcome. Molecular Biology and Genetics Statistics of barcoding coverage: Elminius modestus Public Records: 0 Specimens with Barcodes: 3 Species With Barcodes: 1 Elminius modestus is a species of barnacle in the family Balanidae, native to Australia, Tasmania and New Zealand, but now spread to Britain and the north west coasts of Europe. It reaches a maximum size of about 10 millimetres (0.39 in) in diameter. Distribution and habitat E. modestus originated in Australia and was first seen in British waters, in Chichester Harbour, during the Second World War. It was believed to have arrived on the hulls of ships, or possibly the larval stages travelled in bilge water. It has become very common in southern England and Wales and is spreading northwards, but the spread may be limited by the temperature of the sea. It is found on the upper middle shore and is tolerant of low salinity levels where fresh water enters the sea. It avoids exposed positions. It had reached the Scottish Borders by 1960 and Shetland by 1978. It is found on the Atlantic coasts of Europe from Gibraltar to Germany. E. modestus is a suspension feeder. It has feathery appendages which beat rhythmically to draw plankton and other organic particles into the shell for consumption. Eggs are laid and develop into nauplius larvae which are released into the phytoplankton. These then develop into cyprid larvae which later settle and cement themselves onto a rocky substrate. In the British Isles, E. modestus competes with Semibalanus balanoides, whereas in southern Europe it also competes with Chthamalus spp. It is particularly successful because it grows fast, tolerates reduced salinity, has a lower temperature tolerance than Chthamalus spp and a higher tolerance than Balanus spp. It is also a threat to native species because it reaches maturity in its first season and can produce several broods of larvae per year. It has an extended habitat as it grows both high up the shore and in the neritic zone. - WoRMS (2011). "Elminius modestus". World Register of Marine Species. _URL_ Retrieved August 17, 2011. - John Barrett & C. M. Young (1958). Collins Pocket Guide to the Sea Shore. p. 91. - D. J. Crisp (1958). "The spread of Elminius modestus Darwin in north-west Europe" (PDF). Journal of the Marine Biological Association of the United Kingdom 37 (2): 483–520. doi:10.1017/S0025315400023833. _URL_ - K. Hiscock, S. Hiscock & J. M. Baker (1978). "The occurrence of the barnacle Elminius modestus in Shetland". Journal of the Marine Biological Association of the United Kingdom 58 (3): 627–629. doi:10.1017/S0025315400041278. - H. Barnes & M. Barnes (1966). "Ecological and zoogeographical observations on some of the common intertidal cirripedes of the coasts of the western European mainland in June-September, 1963". In Harold Barnes. Some Contemporary Studies in Marine Science. Allen & Unwin. pp. 83–105. - "Shore life". Encarta Encyclopedia 2005 DVD. - E. Bourget (1987). "Barnacle shells: composition, structure, and growth". In Alan J. Southward. Crustacean Issues 5: Barnacle Biology. pp. 267–285. ISBN 90-6191-628-3. - H. Barnes & Margaret Barnes (1960). "Recent spread and present distribution of the barnacle Elminius modestus Darwin in north-west Europe". Proceedings of the Zoological Society of London 135 (1): 137–145. doi:10.1111/j._PHONE_.1960.tb05836.x. - "Elminius modestus". Marine Advice. Joint Nature Conservation Committee. _URL_ Retrieved August 17, 2011. EOL content is automatically assembled from many different content providers. As a result, from time to time you may find pages on EOL that are confusing. To request an improvement, please leave a comment on the page. Thank you!
2.378989
HuggingFaceFW/fineweb-edu
viscosity or imparting thixotropy in such an amount that does not impair conductivity of the resulting paste composition. Suitable diluting solvents include ethyl cellosolve, butyl cellosolve, ethyl carbitol, butyl carbitol, ethyl cellosolve acetate, etc. The diluting solvent is used for viscosity adjustment to facilitate dispersion in a dispersing apparatus, printing by means of a screen printing machine, etc. A preferred viscosity of the paste composition is from 300 to 1000 poises in a paste state and from 50 to 200 poises for printing. The paste composition of the present invention is prepared usually by means of a dispersing apparatus, e.g., a ball mill, a sand mill, a disper, a Hoover muller, a three-roll mill, a kneader, etc. The conductive paste composition of the present invention is coated or printed by screen printing, transfer printing, or the like technique and then heated to form a cured film. The present invention is now illustrated in greater detail with reference to the following Examples, but it should be understood that the present invention is not deemed to be limited thereto. All the parts and percents are by weight unless otherwise indicated. Fifty parts of tetraglycidyldiaminodiphenylmethane ("Epikote 604" produced by Yuka Shell Epoxy K.K.), 108 parts of a 50% ethyl carbitol solution of a resol type phenol curing agent ("Resitop PS-4101" produced by Gun-ei Kagaku K.K.), and 500 parts of commercially available electrolytic copper powder for industrial use having an average particle size of 5 μm and a loss by reduction of 0.17% by weight were kneaded in a three-roll mill manufactured by Kodaira Seisakusho for 2 minutes to prepare a paste. The resulting paste was coated on a glass epoxy substrate by the use of a screen printing machine (made by Toyo Seimitsu Kogyo K.K.), and the coating film was heated at 150° C. for 5 hours to form a cured film having a thickness of 50 μm. Performance properties of the cured film were evaluated according to the following test methods. The results obtained are shown in Table 1 below. 1) Volume Resistivity Measured with a resistance meter ("LORESTA AP" manufactured by Mitsubishi Petrochemical Co., Ltd.). In Example 1, the volume resistivity of the sample was in the order of 10-5 Ω·cm as shown in Table 1. 2) Moisture Resistance The sample after the resistivity measurement was allowed to stand at 66° C. and 90% RH for 1,000 hours and then its volume resistivity was again measured in the same manner as in 1) above. 3) Adhesion to Resist Determined according to UL 746C. A paste was prepared in the same manner as in Example 1, except for replacing Resitop PS-4101 with Resitop PS-4122 (produced by Gun-ei Kagaku K.K.) as a resol type phenol curing agent. A cured film was formed by using the resulting paste in the same manner as in Example 1. Performance properties of the cured film were evaluated in the same manner as in Example 1, and the results obtained are shown in Table 1. A paste was prepared in the same manner as in Example 1, except for using 100 parts of a bisphenol F type epoxy resin ("Epikote 807" produced by Yuka Shell Epoxy K.K.), 4 parts of dicyandiamide (curing agent), and 400 parts of the same commercially available electrolytic copper powder as that in Example 1. A cured film was formed by using the resulting paste in the same manner as in Example 1. Performance properties of the cured film were evaluated in the same manner as in Example 1, and the results obtained are shown in Table 1. A paste was prepared in the same manner as in Example 1, except for replacing Epikote 604 with 50 parts of tetraglycidyl-m-xylidenediamine ("TETRA D-X" produced by Mitsubishi Gas Chemical Industries, Ltd.). A cured film was formed by using the resulting paste in the same manner as in Example 1. Performance properties of the cured film were evaluated in the same manner as in Example 1, and the results obtained are shown in Table 1. TABLE 1__________________________________________________________________________ Volume Resistivity (Ω · cm) Paste Composition (part) Immediate- After Mois-Example Copper Epoxy Curing ly After ture Resis-No. Powder Resin Agent Curing tance Test Adhesion__________________________________________________________________________Example 1 electrolytic Epikote 604 PS-4101 3 × 10-5 6 × 10-5 5B copper (500) (50) (54)Example 2 electrolytic Epikote 604 PS-4122 6 × 10-5 9 × 10-5 5B copper (500) (50) (54)Example 3 electrolytic TETRA D-X PS-4101 3 × 10-5 5 × 10-5 5B copper (500) (50) (54)Comparative electrolytic Epikote 806 dicyandi- 1 × 10-3 ∞ 5BExample copper (400) (100) amide (4)__________________________________________________________________________ As described and demonstrated above, the conductive paste composition according to the present invention exhibits excellent moisture resistance and adhesion owing to the epoxy matrix resin thereof and provides a cured film of high level of conductivity. While the invention has been described in detail and with reference to specific examples thereof, it will be apparent to one skilled in the art that various changes and modifications can be made therein without departing from the spirit and scope thereof. |Cited Patent||Filing date||Publication date||Applicant||Title| |US2951825 *||Feb 3, 1958||Sep 6, 1960||Union Carbide Corp||Glycidyl derivatives of amino phenols| |US3683044 *||Dec 29, 1970||Aug 8, 1972||Japan Gas Chemical Co||Heat-curable composition comprising polyglycidyl xylylene-diamine and process for preparation of polyglycidyl xylylenediamine| |US3843565 *||Sep 25, 1973||Oct 22, 1974||Mitsubishi Gas Chemical Co||Epoxy resins consisting essentially of n,n,n',n'-tetraglycidyl bisaminomethylcyclo-hexane| |US4201854 *||Sep 28, 1978||May 6, 1980||Ciba-Geigy Corporation||Curable epoxide resin mixtures| |US4410457 *||May 5, 1982||Oct 18, 1983||Sumitomo Chemical Co., Ltd.||Conductive paste| |US4595606 *||Jul 18, 1984||Jun 17, 1986||Rohm And Haas Company||Solderable conductive compositions having high adhesive strength| |US4610810 *||May 17, 1985||Sep 9, 1986||Matsushita Electric Industrial Co., Ltd.||Radiation curable resin, paint or ink vehicle composition comprising said resin and magnetic recording medium or resistor element using said resin| |US4652398 *||Apr 14, 1986||Mar 24, 1987||Stauffer Chemical Company||Rapid curing, thermally stable adhesive composition comprising epoxy resin, polyimide, reactive solvent, and crosslinker| |US4663079 *||Jun 27, 1985||May 5, 1987||Mitsubishi Petrochemical Co., Ltd.||Copper-type conductive coating composition| |US4696764 *||Nov 30, 1984||Sep 29, 1987||Osaka Soda Co., Ltd.||Electrically conductive adhesive composition| |US4732702 *||Jan 20, 1987||Mar 22, 1988||Hitachi Chemical Company, Ltd.||Electroconductive resin paste| |US493701
1.571716
m-a-p/FineFineWeb
Food as a window to our past Redesigned summer course explores society while cooking recipes in the lab A History course with a lab component? Others have tried it, and this summer, for the first time, the Faculty of Arts will offer their own unique take on history with Food Histories (HIST 3760). Dr. Sarah Elvins has reimagined her course to incorporate a weekly kitchen lab component for the first time. "Adding the lab portion to the course allows for a more interactive experience. In the kitchen, we will cook and, of course, eat, as well as explore historic recipes, ingredients, tools, and technology," explains Elvins. Food Histories takes students back in time to investigate the U.S.'s distinct and varied food culture. The course will delve into themes including regional food, Indigenous experiences, immigration, industrialization and agriculture, the rise of advertising and convenience foods, technology, and the evolution of 'home' cooking. Elvins adds, "Food can be so much more than just nutrition, the latest gadget, or the hippest Instagram photo. Food history helps us to explore society: class, gender, technology, communication – even cultural, social, and political implications – through what we ate and how we prepared it." Elvins hopes this unique experiential learning course is both enriching and fun for students. "Studying the past helps us to better understand our experience today. This course educates us on traditional historical topics through the lens of food and now with the weekly kitchen lab, helps make history a hands-on experience." The course will be of interest to a broad audience including those in anthropology, sociology, food science, marketing, entrepreneurship, or human nutrition. HIST 3760, Food Histories will be offered in summer session, June 4-26, Monday to Friday, 10:45 am- 12:45 pm. Course prerequisite is 6 credit hours of 1000 or 2000 History or special permission from the instructor.
2.103682
m-a-p/FineFineWeb
pmid:27824859 32. 32. Pease JB, Hahn MW. More accurate phylogenies inferred from low-recombination regions in the presence of incomplete lineage sorting. Evolution. 2013;67: 2376–2384. pmid:23888858 33. 33. Van Belleghem SM, Rastas P, Papanicolaou A, Martin SH, Arias CF, Supple MA, et al. Complex modular architecture around a simple toolkit of wing pattern genes. Nat Ecol Evol. 2017;1: 0052. pmid:28523290 34. 34. Merrill RM, Gompert Z, Dembeck LM, Kronforst MR, McMillan WO, Jiggins CD. Mate preference across the speciation continuum in a clade of mimetic butterflies. Evolution. 2011;65: 1489–500. pmid:21521198 35. 35. Jiggins CD, Naisbit RE, Coe RL, Mallet J. Reproductive isolation caused by colour pattern mimicry. Nature. 2001;411: 302–5. pmid:_PHONE_ 36. 36. Mérot C, Salazar C, Merrill RM, Jiggins C, Joron M. What shapes the continuum of reproductive isolation? Lessons from Heliconius butterflies. Proc R Soc B Biol Sci. 2017;284: 20170335. 37. 37. Jiggins C. Ecological speciation in mimetic butterflies. Bioscience. 2008;58: 541–548. 38. 38. Merot C, Frerot B, Leppik E, Joron M. Beyond magic traits: Multimodal mating cues in Heliconius butterflies. Evolution. 2015; 2891–2904. pmid:26513426 39. 39. Darragh K, Vanjari S, Mann F, Gonzalez-Rojas MF, Morrison CR, Salazar C, et al. Male sex pheromone components in Heliconius butterflies released by the androconia affect female choice. PeerJ. 2017;5: e3953. pmid:29134139 40. 40. Jiggins CD, Estrada C, Rodrigues a. Mimicry and the evolution of premating isolation in Heliconius melpomene Linnaeus. J Evol Biol. 2004;17: 680–91. pmid:_PHONE_ 41. 41. Naisbit RE, Jiggins CD, Mallet J. Disruptive sexual selection against hybrids contributes to speciation between Heliconius cydno and Heliconius melpomene. Proc Biol Sci. 2001;268: 1849–54. pmid:_PHONE_ 42. 42. Jiggins CD, Linares M, Naisbit RE, Salazar C, Yang ZH, Mallet J. Sex-linked hybrid sterility in a butterfly. Evolution. 2001;55: 1631–1638. pmid:_PHONE_ 43. 43. Merrill RM, Wallbank RWR, Bull V, Salazar PC a, Mallet J, Stevens M, et al. Disruptive ecological selection on a mating cue. Proc Biol Sci. 2012;279: 4907–13. pmid:23075843 44. 44. Sánchez AP, Pardo-Diaz C, Enciso-Romero J, Muñoz A, Jiggins CD, Salazar C, et al. An introgressed wing pattern acts as a mating cue. Evolution. 2015;69: 1619–1629. pmid:25930106 45. 45. Kronforst MRR, Hansen MEB, Crawford NGG, Gallant JRR, Zhang W, Kulathinal RJJ, et al. Hybridization reveals the evolving genomic architecture of speciation. Cell Rep. 2013;5: 666–77. pmid:24183670 46. 46. Lohse K, Chmelik M, Martin SH, Barton NH. Efficient strategies for calculating blockwise likelihoods under the coalescent. Genetics. 2016;202: 775–786. pmid:26715666 47. 47. Pardo-Diaz C, Salazar C, Baxter SW, Merot C, Figueiredo-Ready W, Joron M, et al. Adaptive introgression across species boundaries in Heliconius butterflies. PLoS Genet. 2012;8: e1002752. pmid:22737081 48. 48. Heliconius Genome Consortium T. Butterfly genome reveals promiscuous exchange of mimicry adaptations among species. Nature. 2012;487: 94–8. pmid:22722851 49. 49. Enciso-Romero J, Pardo-Díaz C, Martin SH, Arias CF, Linares M, McMillan WO, et al. Evolution of novel mimicry rings facilitated by adaptive introgression in tropical butterflies. Mol Ecol. 2017;26: 5160–5172. pmid:28777894 50. 50. Davey JW, Barker SL, Rastas PM, Pinharanda A, Martin SH, Durbin R, et al. No evidence for maintenance of a sympatric Heliconius species barrier by chromosomal inversions. Evol Lett. 2017;1: 138–154. pmid:30283645 51. 51. Nadeau NJ, Martin SH, Kozak KM, Salazar C, Dasmahapatra KK, Davey JW, et al. Genome-wide patterns of divergence and gene flow across a butterfly radiation. Mol Ecol. 2013;22: 814–26. pmid:22924870 52. 52. Martin SH, Moest M, Palmer WJ, Salazar C, McMillan WO, Jiggins FM, et al. Natural selection and genetic diversity in the butterfly Heliconius melpomene. Genetics. 2016;203: 525–541. pmid:27017626 53. 53. Rosser N, Phillimore AB, Huertas B, Willmott KR, Mallet J. Testing historical explanations for gradients in species richness in heliconiine butterflies of tropical America. Biol J Linn Soc. 2012;105: 479–497. 54. 54. Becker RA, Wilks AR, Brownrigg R, Minka TP, Deckmyn A. maps: Draw Geographical Maps. 2018. _URL_ 55. 55. Martin SH (2019). Data from: Recombination rate variation shapes barriers to introgression across butterfly genomes. Dryad Digital Repository. Openly available via _URL_ 56. 56. Martin SH, Van Belleghem SM. Exploring evolutionary relationships across the genome using topology weighting. Genetics. 2017;206. pmid:28341652 57. 57. Martin SH, Eriksson A, Kozak KM, Manica A, Jiggins CD. Speciation in Heliconius Butterflies: Minimal Contact Followed by Millions of Generations of Hybridisation. bioRxiv. 2015; 015800. 58. 58. Kozak KM, Wahlberg N, Neild AFE, Dasmahapatra KK, Mallet J, Jiggins CD. Multilocus species trees show the recent adaptive radiation of the mimetic heliconius butterflies. Syst Biol. 2015;64: 505–524. pmid:25634098 59. 59. Estrada C, Jiggins CD. Patterns of pollen feeding and habitat preference among Heliconius species. Ecol Entomol. 2002;27: 448–456. 60. 60. Giraldo N, Salazar C, Jiggins CD, Bermingham E, Linares M. Two sisters in the same dress: Heliconius cryptic species. BMC Evol Biol. 2008;8: 324. pmid:_PHONE_ 61. 61. Ahola V, Lehtonen R, Somervuo P, Salmela L, Koskinen P, Rastas P, et al. The Glanville fritillary genome retains an ancient karyotype and reveals selective chromosomal fusions in Lepidoptera. Nat Commun. 2014;5: 1–9. pmid:25189940 62. 62. Davey JW, Chouteau M, Barker SL, Maroja L, Baxter SW, Simpson F, et al. Major Improvements to the Heliconius melpomene Genome Assembly Used to Confirm 10 Chromosome Fusion Events in 6 Million Years of Butterfly Evolution. G3. 2016;6: 695–708. pmid:26772750 63. 63. Baxter SW, Nadeau NJ, Maroja LS, Wilkinson P, Counterman BA, Beltran M, et al. Genomic Hotspots for Adaptation: The Population Müllerian Mimicry in the Heliconius Genetics of Mu melpomene Clade. Nachman MW, editor. Genetics. 2010;6: e1000794. pmid:20140188 64. 64. Wallbank RWR, Baxter SW, Pardo-Diaz C, Hanly JJ, Martin SH, Mallet J, et al. Evolutionary Novelty in a Butterfly Wing Pattern through Enhancer Shuffling. PLoS Biol. 2016;14: 1–16. pmid:26771987 65. 65. Nadeau NJ, Pardo-diaz C, Whibley A,
1.055364
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
In the vast world of seafood, mussels stand as a culinary gem, offering a taste that is both distinctive and nuanced. Picture a delightful balance of sweetness, a gentle brininess from their saltwater habitat, and a subtle umami essence. These bivalve mollusks, reminiscent of clams and oysters but with their own unique character, invite us into a sensory journey from the ocean to our plates. Mussels, whether enjoyed on their own or as part of a flavorful seafood dish, tantalize the taste buds with their tender texture and marine-inspired palette. In this exploration, we delve into the exquisite taste profile of mussels, examining the factors that influence their flavor, the art of cooking them to perfection, and the myriad ways they contribute to a healthy and delectable dining experience. Join us on a culinary adventure as we unravel the question: What do mussels taste like? Table of Contents What are Mussels? Before delving into the nuances of the taste and culinary applications of mussels, it's crucial to understand what mussels are. Mussels are bivalve mollusks belonging to the Mytilidae family. These shellfish are characterized by their dark, oblong shells, which typically house the tender and flavorful meat inside. Found in both freshwater and saltwater environments, mussels play a vital ecological role as filter feeders, purifying the water by extracting microscopic particles. Mussels are known for their strong byssal threads, which they use to anchor themselves to rocks or other substrates in their natural habitat. These threads not only keep them in place against the constant motion of the waves but also contribute to their identity as resilient and adaptable creatures. Read: What Does Swordfish Taste Like? Explained With Detailed Guide. Types of Mussels Now that we have laid the foundation by understanding what mussels are, let's further explore the diverse array of mussel varieties that grace our oceans and coastal regions. Each type of mussel comes with its own unique characteristics, contributing to the rich tapestry of marine life and global culinary traditions. 1. Blue Mussel (Mytilus edulis): Description: Recognizable by their dark blue-black shells, Blue Mussels are prevalent in both North American and European waters. They thrive in intertidal zones, clinging to rocks and substrates along the shoreline. Blue Mussels are celebrated for their tender meat and are a staple in various seafood dishes worldwide. 2. Mediterranean Mussel (Mytilus galloprovincialis): Description: Distinguished by their smaller size and brownish-green shells, Mediterranean Mussels are commonly found in the Mediterranean Sea and along the Atlantic coast of Portugal and Spain. They prefer warmer waters and are known for their robust flavor, making them a popular choice in Mediterranean cuisine. 3. Green-Lipped Mussel (Perna canaliculus): Description: Native to New Zealand, the Green-Lipped Mussel stands out with its vibrant green shell edges. These mussels are larger in size, offering juicy, plump meat. The taste is often described as sweet and delicate, and they are a prized ingredient in both traditional and modern New Zealand dishes. 4. California Mussel (Mytilus californianus): Description: Inhabiting the rocky shores of the Pacific coast, the California Mussel is recognized for its robust flavor. With a combination of brininess and a slightly nutty undertone, these mussels contribute a unique taste to coastal cuisine. Their firm texture adds substance to a variety of dishes. 5. Asian Green Mussel (Perna viridis): Description: Commonly found in coastal waters of Asia, the Asian Green Mussel is characterized by its dark green shell. These mussels have a sweet and briny taste, making them a popular choice in Asian cuisine. They are often featured in dishes such as curries and stir-fries. 6. Chilean Mussel (Mytilus chilensis): Description: Hailing from the southern regions of South America, the Chilean Mussel is known for its large size and bold flavor. With a meaty texture and a taste that balances sweetness with a touch of minerality, these mussels are frequently used in Chilean seafood dishes. Exploring the different types of mussels allows us to appreciate the biodiversity of these bivalve mollusks. Each variety not only adapts to its specific habitat but also brings a unique set of flavors and textures to the world of culinary exploration. Now that we've established the foundation of what mussels are and the types available, let's dive into the heart of the matter – the taste of mussels. Read: What Does Conch Taste Like? Side Effects & More. What Do Mussels Taste Like? Mussels, with their slightly sweet and delicate flavor, offer a culinary experience that can be likened to clams or oysters. The subtle brininess they possess is a testament to their exposure to saltwater, creating a refreshing taste that mirrors the essence of the ocean. These bivalve mollusks, while having a mild base flavor, are far from bland. Mussels have a unique ability to absorb the flavors of the ingredients they are cooked with, making them a versatile canvas for culinary creativity. Spices and herbs, when introduced to the cooking process, enhance the inherent taste of mussels, creating a symphony of flavors that dance on the palate. The taste of mussels can vary depending on their kind, but a general consensus describes them as mild with a meaty undertone that exudes umami. Many compare the taste of mussels to their shellfish counterparts like clams and oysters. Unlike oysters, mussels are less slimy and briny, offering a cleaner and more approachable flavor profile. While mussels don't possess the tenderness of clams, they make up for it with a satisfying meatiness that adds substance to their taste. The texture of mussels, when properly cooked, strikes a balance between firmness and tenderness, creating a delightful mouthfeel. What truly distinguishes the taste of mussels is the unmistakable essence of the ocean they carry. Mussels taste undeniably salty, a direct reflection of their marine habitat. This briny quality contributes to the overall character of the dish, reminding the diner of the sea's influence on this delectable mollusk. Some connoisseurs of seafood also detect a subtle undertone of mushrooms in the flavor profile of mussels, adding a layer of complexity to their taste. This light sweetness, reminiscent of mushrooms, further enhances the culinary intrigue surrounding mussels. It's essential to recognize that while mussels, clams, and oysters share similarities as shellfish, each possesses its own distinct characteristics. The contrasting differences become evident upon closer examination and tasting, emphasizing the unique qualities that make mussels a standout in the world of seafood. What Do Other Mussels Taste Like? While we've delved into the nuanced taste of the Blue Mussel (Mytilus edulis) and the Mediterranean Mussel (Mytilus galloprovincialis), it's important to acknowledge that the world of mussels extends beyond these commonly consumed varieties. Other mussels, each with its own unique characteristics, contribute to the rich tapestry of flavors in the culinary realm. 1. Green-Lipped Mussel (Perna canaliculus): Characteristics: Native to New Zealand, the Green-Lipped Mussel is renowned for its vibrant green shell edges and plump, juicy meat. The taste of the Green-Lipped Mussel is often described as sweet and delicate, with a pronounced oceanic freshness. This mussel variety is larger than its counterparts, offering a generous bite that seafood enthusiasts appreciate. 2. California Mussel (Mytilus californianus): Characteristics: Inhabiting the rocky shores of the Pacific coast, the California Mussel boasts a robust flavor profile. Its taste is characterized by a combination of brininess and a slightly nutty undertone. These mussels are often enjoyed for their firm texture and the savory richness they lend to various dishes. 3. Asian Green Mussel (Perna viridis): Characteristics: Commonly found in the coastal waters of Asia, the Asian Green Mussel has a distinctive dark green shell and a taste that is both sweet and briny. This variety is often featured in Asian cuisine, contributing its unique flavor to dishes like curries and stir-fries. 4. Chilean Mussel (Mytilus chilensis): Characteristics: Hailing from the southern regions of South America, the Chilean Mussel is known for its large size and bold flavor. These mussels have a meaty texture and a taste that balances sweetness with a touch of minerality. They are frequently used in Chilean seafood dishes, showcasing the diverse applications of mussels in global gastronomy. Understanding the taste profiles of these different mussel varieties provides a broader perspective on the culinary possibilities that this shellfish offers. The diverse marine ecosystems in which these mussels thrive contribute to the unique flavors they develop, making each variety a culinary adventure waiting to be explored. As we journey through the world of mussels, it becomes evident that the taste experience extends far beyond a single characteristic. The subtle nuances, influenced by factors such as habitat, size, and local cuisine, highlight the versatility and complexity that make mussels a beloved ingredient in kitchens around the world. Do Mussels Taste Like Oysters? The comparison between mussels and oysters often arises due to their shared
2.080064
m-a-p/FineFineWeb
injection. After 28 days of T1DM, KO mice exhibited a similar reduction in body weight and plasma insulin levels and a comparable increase in glycemia (WT: 384 ± 20 vs. Ko: 432 ± 29 mg/dL, cholesterol and triglycerides, as WT mice. T1DM increased PTP1B expression and impaired endothelial NO-dependent relaxation, in mouse aorta. PTP1B deletion did not affect baseline endothelial function, but preserved endothelium-dependent relaxation, in T1DM mice. NO synthase inhibition with L-NAME abolished endothelial relaxation in control and T1DM WT mice, whereas L-NAME and the cyclooxygenases inhibitor indomethacin were required to abolish endothelium relaxation in T1DM KO mice. PTP1B deletion increased COX-2 expression and PGI2 levels, in mouse aorta and plasma respectively, in T1DM mice. In parallel, simulation of diabetic conditions increased PTP1B expression and knockdown of PTP1B increased COX-2 but not COX-1 expression, in primary human aortic endothelial cells. Taken together these data indicate that deletion of PTP1B protected endothelial function by compensating the reduction in NO bioavailability by increasing COX-2-mediated release of the vasodilator prostanoid PGI2, in T1DM mice. 20. Detection of phosphatase activity in aquatic and terrestrial cyanobacterial strains Directory of Open Access Journals (Sweden) Babić Olivera B. 2013-01-01 Full Text Available Cyanobacteria, as highly adaptable microorganisms, are characterized by an ability to survive in different environmental conditions, in which a significant role belongs to their enzymes. Phosphatases are enzymes produced by algae in relatively large quantities in response to a low orthophosphate concentration and their activity is significantly correlated with their primary production. The activity of these enzymes was investigated in 11 cyanobacterial strains in order to determine enzyme synthesis depending on taxonomic and ecological group of cyanobacteria. The study was conducted with 4 terrestrial cyanobacterial strains, which belong to Nostoc and Anabaena genera, and 7 filamentous water cyanobacteria of Nostoc, Oscillatoria, Phormidium and Microcystis genera. The obtained results showed that the activity of acid and alkaline phosphatases strongly depended on cyanobacterial strain and the environment from which the strain originated. Higher activity of alkaline phosphatases, ranging from 3.64 to 85.14 μmolpNP/s/dm3, was recorded in terrestrial strains compared to the studied water strains (1.11-5.96 μmolpNP/s/dm3. The activity of acid phosphatases was higher in most tested water strains (1.67-6.28 μmolpNP/s/dm3 compared to the activity of alkaline phosphatases (1.11-5.96 μmolpNP/s/dm3. Comparing enzyme activity of nitrogen fixing and non-nitrogen fixing cyanobacteria, it was found that most nitrogen fixing strains had a higher activity of alkaline phosphatases. The data obtained in this work indicate that activity of phosphatases is a strain specific property. The results further suggest that synthesis and activity of phosphatases depended on eco-physiological characteristics of the examined cyanobacterial strains. This can be of great importance for the further study of enzymes and mechanisms of their activity as a part of cyanobacterial survival strategy in environments with extreme conditions. [Projekat Ministarstva nauke Republike 1. Interaction of Uranium with Bacterial Cell Surfaces: Inferences from Phosphatase-Mediated Uranium Precipitation Science.gov (United States) Kulkarni, Sayali; Misra, Chitra Seetharam; Gupta, Alka; Ballal, Anand 2016-01-01 ABSTRACT Deinococcus radiodurans and Escherichia coli expressing either PhoN, a periplasmic acid phosphatase, or PhoK, an extracellular alkaline phosphatase, were evaluated for uranium (U) bioprecipitation under two specific geochemical conditions (GCs): (i) a carbonate-deficient condition at near-neutral pH (GC1), and (ii) a carbonate-abundant condition at alkaline pH (GC2). Transmission electron microscopy revealed that recombinant cells expressing PhoN/PhoK formed cell-associated uranyl phosphate precipitate under GC1, whereas the same cells displayed extracellular precipitation under GC2. These results implied that the cell-bound or extracellular location of the precipitate was governed by the uranyl species prevalent at that particular GC, rather than the location of phosphatase. MINTEQ modeling predicted the formation of predominantly positively charged uranium hydroxide ions under GC1 and negatively charged uranyl carbonate-hydroxide complexes under GC2. Both microbes adsorbed 6- to 10-fold more U under GC1 than under GC2, suggesting that higher biosorption of U to the bacterial cell surface under GC1 may lead to cell-associated U precipitation. In contrast, at alkaline pH and in the presence of excess carbonate under GC2, poor biosorption of negatively charged uranyl carbonate complexes on the cell surface might have resulted in extracellular precipitation. The toxicity of U observed under GC1 being higher than that under GC2 could also be attributed to the preferential adsorption of U on cell surfaces under GC1. This work provides a vivid description of the interaction of U complexes with bacterial cells. The findings have implications for the toxicity of various U species and for developing biological aqueous effluent waste treatment strategies. IMPORTANCE The present study provides illustrative insights into the interaction of uranium (U) complexes with recombinant bacterial cells overexpressing phosphatases. This work demonstrates the effects of aqueous 2. Improving the expression of recombinant proteins in E. coli BL21 (DE3) under acetate stress: an alkaline pH shift approach. Science.gov (United States) Wang, Hengwei; Wang, Fengqing; Wang, Wei; Yao, Xueling; Wei, Dongzhi; Cheng, Hairong; Deng, Zixin 2014-01-01 Excess acetate has long been an issue for the production of recombinant proteins in E. coli cells. Recently, improvements in acetate tolerance have been achieved through the use of genetic strategies and medium supplementation with certain amino acids and pyrimidines. The aim of our study was to evaluate an alternative to improve the acetate tolerance of E. coli BL21 (DE3), a popular strain used to express recombinant proteins. In this work we reported the cultivation of BL21 (DE3) in complex media containing acetate at high concentrations. In the presence of 300 mM acetate, compared with pH 6.5, pH 7.5 improved cell growth by approximately 71%, reduced intracellular acetate by approximately 50%, and restored the expression of glutathione S-transferase (GST), green fluorescent protein (GFP) and cytochrome P450 monooxygenase (CYP). Further experiments showed that alkaline pHs up to 8.5 had little inhibition in the expression of GST, GFP and CYP. In addition, the detrimental effect of acetate on the reduction of 3-(4,5-dimethylthiazol-2-yl)-2,5-diphenyl tetrazolium bromide (MTT) by the cell membrane, an index of cellular metabolic capacity, was substantially alleviated by a shift to alkaline pH values of 7.5-8.0. Thus, we suggest an approach of cultivating E. coli BL21 (DE3) at pH 8.0 ± 0.5 to minimize the effects caused by acetate stress. The proposed strategy of an alkaline pH shift is a simple approach to solving similar bioprocessing problems in the production of biofuels and biochemicals from sugars. 3. Simplified preparation of a phosphatase inhibitor and further studies of its action. Science.gov (United States) Coburn, S P; Schaltenbrand, W E 1978-05-01 1-Pyrrolidinecarbothioic acid (2-pyridylmethylene) hydrazide chelates Zn2+ but not Mg2+. This compound is about twice as effective as EDTA for inhibiting alkaline phosphatase from calf mucosa, and approx. 1000-fold more effective than EDTA for inhibiting acid phosphatase from wheat germ. The compound did not inhibit pyridoxine kinase activity in human leucocytes at the highest concentration tested (33 micron). Therefore it may be a useful tool for either examining or eliminating the effects of phosphatases in complex enzyme systems. 4. The effect of extracellular alkalinization on
1.180907
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
Australasian Science: Australia's authority on science since 1938 Which Pregnant Women Are at Risk? Credit: Olesia Bilkei/Adobe Credit: Olesia Bilkei/Adobe By Claire Roberts & Tina Bianco-Miotto Women who are pregnant for the first time are largely unaware that the four main complications of pregnancy – pre­eclampsia, preterm birth, intrauterine growth restriction and gestational diabetes – occur in one-quarter of first pregnancies. Worryingly, the number of these pregnancy complications is rising, and they contribute to poor health outcomes for mothers and babies in both the short and long term. Currently there are no effective screening tools in clinical practice that can accurately predict any complications of pregnancy in women who are pregnant for the first time. While there are models to predict some of these complications in women who have previously given birth, they rely heavily on previous pregnancy history and their use is still confined to a small number of UK hospitals. Furthermore, their accuracy is modest at best. However, our research group has used data from nearly 3000 pregnant women in Adelaide and Auckland, their partners and babies to develop robust models to predict risk for each of the four pregnancy complications with low false positives and negatives. These will be offered as a genetic test, for which a small blood or saliva sample is required from both the pregnant woman and her partner at 12–15 weeks gestation. We have already identified specific gene variants in both the pregnant woman and her partner that, when combined with a number of clinical, family history, dietary and lifestyle factors, predict risk. Some of these factors are highly modifiable. In this article we will reveal what some of these factors are and offer strategies to reduce women's risk. Risk Factors We are all predisposed to some sort of disease, such as heart disease, high blood pressure or some cancers. These tend to run in families, and this implicates genetic variants in their aetiology. We have identified a number of gene variants in both pregnant women and their partners that contribute to the risk of pregnancy complications. Gene variants that associate with many diseases, including pregnancy complications, are often said to have low penetrance. This simply means that not all people who carry these variants actually get the disease. Each person carries a unique combination of millions of gene variants, so we all have protective and deleterious gene variants that interact as a suite. Individual gene variants can affect the expression of the gene that they encode but they can also have significant effects on the expression of other genes located some distance away within the genome. We have also found that the health of the parents of a pregnant woman can also indicate her risk for pregnancy complications. For example, a woman whose father has chronic high blood pressure is at a greater risk herself of developing pre­eclampsia, the most serious hypertensive disorder of pregnancy. Family history of miscarriage and pregnancy complications in the woman's mother or sisters can also indicate an increased risk. These suggest that there are genetic factors common to both cardiovascular and metabolic diseases and pregnancy complications. Indeed, we have found this to be the case. Genes are also known to interact with the environment in both health and disease. This means that an individual may harbour a number of gene variants that pre­dispose them to a particular disease but it is only concomitant exposure to certain environmental factors, such as a poor diet and physical inactivity, that reveals a vulnerability to that disease. Gene expression can directly be influenced by nutrition. For example, zinc is a structural component of hundreds of enzymes and transcription factors that regulate gene expression, and its deficiency may compromise the function of a number of biological processes. Environmental exposures occur in the context of the foetal, and hence placental, genome. The foetus and placenta cross-communicate with the mother, and both are at the mercy of her past and present environmental exposures, including her diet. Furthermore, the placenta orchestrates the mother's response to pregnancy by secreting a number of hormones and growth factors into her circulation. These are determined by the foetal and placental genome, and also respond to her health status and exposures. Epigenetic Risks Although a person's DNA code is identical in every cell, differences in the DNA's epigenetic state can determine cell and tissue type. For instance, a liver cell has different epigenetic markers than a heart cell in the same person. There is evidence environmental factors can repress or amplify gene expression by altering a gene's epi­genetic state. For example, active smoking alters DNA methylation in a number of genes, including tumour suppressors and genes associated with coronary artery disease. Likewise dietary and lifestyle factors can alter the epigenetic state in a variety of human tissues, and lead to the unmasking of disease states that increase the risk of complications during pregnancy. Even the sex of the individual adds to the epigenetic web, with males and females responding differently to such exposures. We have even found that sex differences start in utero so that maternal health, foetal growth and newborn outcomes are different in women carrying a male or a female foetus. Other Factors Folic acid supplementation in the month prior to conception and during the first trimester of pregnancy can prevent neural tube defects such as spina bifida in babies. We have found that first trimester folic acid supplementation also protects women from developing preeclampsia, delivering preterm and having a small-for-gestational-age baby. This suggests that folic acid supplementation may be beneficial to the placenta and how the mother adapts to pregnancy, as well as to the foetus. This has considerable biological plausibility, since folate is the pivotal factor in one-carbon metabolism, and thereby production of S-adenosyl methionine, the universal donor of methyl groups in DNA methylation. Therefore, folic acid supplementation can contribute to the appropriate epigenetic state for correct gene expression and healthy development. We have also found that women can significantly reduce their risk of developing pregnancy complications if they eat three or more serves of fruit and/or one or more serves of green vegetables each day in the month prior to conception and during the first trimester. Fruit and vegetables contain folate as well as other micronutrients and fibre that together may help to improve the outcome of pregnancy. Maternal smoking damages the placenta and impairs foetal growth, and is associated with foetal growth restriction and preterm birth. Maternal illicit drug use is similarly damaging. In fact, we have shown that continued use of marijuana, independent of cigarette smoking, at 20 weeks gestation is strongly associated with preterm birth. Maternal overweight and obesity are increasingly prevalent, with more than 50% of Australian women commencing their pregnancy in these categories. A body mass index above 30 indicates obesity and significantly increases the risk for most pregnancy complications as well as long-term health issues for offspring. Obese women are more likely to deliver a baby with birth defects than lean women, and their children are more likely to be obese and develop type 2 diabetes early in life. Screening for Those at Risk The screening tools we have developed to predict the risk of pregnancy complications include many of these factors in different combinations for each disease. Importantly, they also point to potential new interventions. Despite each of us being at genetic risk, our genes are just a part of the story. Although we cannot change our genomes, we can change our epigenomes and modify the effects of our genes. Fortunately, many of the factors that contribute to pregnancy complications are modifiable. Ideally, pregnancies should be planned ahead and women should achieve a normal weight, have a good diet with recommended supplementation of micronutrients, and engage in moderate exercise both before and during pregnancy. Pre-conception planning can help to minimise the risks to maternal and infant health. One can't help but think that mum was right when she told us to eat our fruit and vegetables. Claire Roberts is a NHMRC Senior Research Fellow who leads the Pregnancy and Birth Theme at The University of Adelaide, Robinson Research Institute. Tina Bianco-Miotto is a lecturer at The University of Adelaide's School of Agriculture, Food and Wine.
2.157077
Zyphra/Zyda-2
The parents and the teachers who need to communicate questions or requests several times in a row often think that the children have hearing problems or they disrespect the adult. Similar conclusions are drawn when the children areunable to react to complex orders or they work slower than their peers. Such children are often labelled as "naughty", which is not true. The prpblem above is illustrated in the text below. Learning and communicating are processes that can seem easy as they are natural for most of us. But for one in ten people, it's a daily battle. The ability to properly process sensory information is compromised. It is said in this case, that listening is disrupted. In the processing of information, the ear plays a fundamental role since it is the main gateway to our brain. It picks up the surrounding sounds and those of our own voice, converts them into electrical stimuli, and transmits them to the brain which analyzes them. More surprisingly, it also captures our movements and is heavily involved in coordination, balance and rhythm. It therefore has a vital function in daily life, whether through talk, to argue, to follow instructions, read, learn, memorize or simply to move. When communication between the ear and the brain is blurred, our ability to interact with the outside world is compromised. The result is often a loss of self-confidence that will enhance in turn the difficulties of communication and learning. The Tomatis® Method is an approach complementary to a therapeutic, pedagogical or medical accompaniment. During the first meeting you will be given an individualized assessment adapted specifically to your problem, followed by an interview and data analysis. Tomatis Listening Test Tomatis Method therapy consists in the intensive neurosensory stimulation. Our Tomatis® professional will create an exclusive program tailored to your needs. During your passive listening sessions, you will use Tomatis devices. The duration, the rhythm and the number of sessions are determined by the Tomatis® professional. In general, the program includes sessions of up to 2 hours which are conducted over a time-frame of at least 3 or 4, 10-13-day periods. Between each period, you take a break of several weeks. These pauses are called "integration phases". They will be followed by tests and intermediate interviews before resuming a new listening session. Qualified professionals may also offer active listening sessions for you. These sessions include reading aloud or verbal repetition, during which your voice is instantly modified and retransmitted within the specific parameters of the Tomatis® effect. Directly and in real-time, you hear your voice corrected, and as a result your voice is unconsciously modified. This exercise is not mandatory but strongly recommended when possible, and is proposed after an initial program of passive listening. Active listening sessions are not systematic and depend on your profile and issues. Exit assessment and summary Once the listening sessions are finished, you are given an exit assessment, which allows us to measure your progress and to determine if further sessions are recommended. The results obtained with the Tomatis® Method are long-lasting, even permanent, and consequently it is often not necessary to extend the sessions beyond the initial course (though sometimes, for certain problems, extending the program may be advised). Listening programme - neurosensory stimulation We have to remember that listening does not equal hearing. Listening is the ability to use our senses consciously in order to learn and communicate. Thus listening is a high level cognitive function, which is closely related to the emotions management. One can listen well, but they can hear poorly. Tomatis Method stimulates the brain. Listen well and change your life! Dr Tomatis discovered the ears play a vital role in a range of functions – beyond simply helping us to hear – including: - Stress and anxiety - Motor skills - Emotion regulation This led him to develop a non-invasive auditory therapy technique to stimulate and enhance the function of the ears which, in many cases, led to improvement of these related functions. How the Tomatis® Method works Using specialised auditory equipment known as the Electronic Ear, electronically modified music and speech is played at specific frequencies to stimulate certain parts of the ear. As these areas respond to the stimulation, so too do the corresponding areas in the brain and nervous system. In turn, this can lead to an improvement in auditory, visual and motor functions which had not previously been receiving the correct level of stimulation and signals from the brain. The treatment typically consists of 2-3 x 15 day phases with a 4-6 week break between, but will largely depend on the nature of the problem and how the patient responds. Some people can see significant improvement in less than a week. What might the Tomatis® Method help with? A range of learning difficulties can be connected to auditory function disorders. Understanding how the ear can affect language skills, reading ability, speech and motor skills, it's not surprising that following treatment many patients have reported: - Improved focus - Enhanced reading/comprehension skills - Reduced stress/anxiety levels - Less "brain fog" - Increased self-awareness - Improved memory and attention The Tomatis® Method has also been found to provide benefits in a number of auditory and learning disorders, including: - Speech problems - Auditory Processing Disorder - Dyslexia, Learning Difficulties - Attention, Memory Our facility in Warsaw is purpose-built to meet the unique needs of our clients, their siblings and families as we 'help children overcome challenges in movement, play, speech, language, learning and behaviour', whilst also meeting their needs for safety, cleanliness and comfort. SENSORIUM facilities provide a professional environment in which our team of highly-trained practitioners can excel at what they do best – helping children develop their skills and achieve their optimum potential. The overriding focus of our facilities is to offer an environment that is physically and emotionally safe. Our therapeutic environment provides a range of opportunities for sensory input while promoting physical development, encouraging social relationships and enhancing independence and self-confidence. Areas of application Attention is the ability to select and maintain in consciousness an external event or thought. It corresponds to a general state of arousal and alertness that enables the nervous system to be receptive to any form of information reaching it. Attention difficulties will concern children or adults who cannot concentrate for long enough periods of work to be done, even if it requires little intellectual effort and is routine or familiar (eg, revision work or 'copy). A particular modality of attention is the ability to be able to divide one's attention between multiple sources of information or tasks. This attentional modality is very expensive cognitively. When it fails, it can be a source of difficulty and considerable suffering for the child in a school situation. In addition to its general action of cerebral stimulation (known as cortical arousal), the Tomatis® Method can have a very positive effect on selective attention. Indeed, the Tomatis® Method relies on the electronic Gating® system that causes a sound contrast, the effect of which constantly and naturally surprises the brain. The goal is to help the brain develop automatic change detection mechanisms. Once these mechanisms are assimilated, the individual can more easily select relevant information and decide to reject or discriminate side events. For this reason, the Tomatis® Method is a complementary approach to support children and adults with attention deficit disorder with or without hyperactivity (ADD / ADHD). Learning is all the mechanisms by which we acquire automation or knowledge. It is the ability to change consistently and persistently one's behavior in response to information or an external event. Learning depends on the ability of an individual to change his behavior based on experience. There are many learning methods, such as learning by reward, by failure, by imitation, by repetition, by creativity... Whatever the method, learning involves the brain's ability to constantly reorganize the neural network to incorporate new behavior and to facilitate its implementation. We speak of "brain plasticity." Thus, due to brain plasticity, when the same event repeats, behavior in response is faster, more accurate and more intuitive. With sound contrasts arranged to surprise the brain naturally, the Tomatis® Method acts on this brain plasticity. Indeed, it helps the brain develop detection mechanisms and analysis of change. Thus, the Tomatis® Method is a pedagogy that aims to train the individual to pay particular attention to external and new events. It facilitates the integration of new information and allows the student to improve his general and academic skills. This is why the Tomatis® Method is currently offered in hundreds of public and private schools in addition to a traditional education. During his schooling, the child learns primarily from mistakes. While it is normal and necessary to make mistakes to grow, it is the persistence of certain targeted and systematic errors that should lead to talk about learning disabilities, mostly grouped under the heading of 'DYS'. These learning problems are not related with intelligence but characterized by a difficulty for the child to properly operate the neural network or certain brain areas. The purpose of the Tomatis® Method is to enable the child to establish compensatory circuits. In other words, it is not possible to cure a learning disability, but it is possible to bypass or compensate it, allowing the brain to exploit other neural networks. For example, for learning disabilities linked to language, the Tomatis® Method facilitates the decoding and analysis of sound and can be an invaluable support in cases of dyslexia and in addition to speech therapy. The accuracy with which we perceive and analyze a sound, affects our faculties to reproduce it. A quality voice transmission also requires good self-listening, that is to say, it requires the ability to listen to our own voice to measure its accuracy, quality and intensity. The sound of our voice is mainly transmitted to the ear by the vibration of the bones of our heads: this is the "bone conduction". It is opposed to the "air conduction", that
2.277713
openbmb/Ultra-FineWeb
Reference : PRoNTo: Pattern Recognition for Neuroimaging Toolbox Scientific journals : Article Engineering, computing & technology : Multidisciplinary, general & others Human health sciences : Multidisciplinary, general & others _URL_ PRoNTo: Pattern Recognition for Neuroimaging Toolbox English Schrouff, Jessica mailto [Université de Liège - ULg > > Centre de recherches du cyclotron >] Rosa, Maria Joao [] Rondina, Jane [] Marquand, Andre [] Chu, Carlton [] Ashburner, John [] Phillips, Christophe [Université de Liège - ULg > > Centre de recherches du cyclotron >] Richiardi, Jonas [] Mourão-Miranda, Janaina [] Feb-2013 Neuroinformatics 11 3 319-337 Yes (verified by ORBi) International _PHONE_ _PHONE_ [en] machine learning ; neuroimaging ; toolbox ; image analysis ; multivariate pattern analysis [en] In the past years, mass univariate statistical analyses of neuroimaging data have been complemented by the use of multivariate pattern analyses, especially based on machine learning models. While these allow an increased sensitivity for the detection of spatially distributed effects compared to univariate techniques, they lack an established and accessible software framework. The goal of this work was to build a toolbox comprising all the necessary functionalities formultivariate analyses of neuroimaging data, based on machine learning models. The "Pattern Recognition for Neuroimaging Toolbox" (PRoNTo) is open-source, cross-platform, MATLAB-based and SPM compatible, therefore being suitable for both cognitive and clinical neuroscience research. In addition, it is designed to facilitate novel contributions from developers, aiming to improve the interaction between the neuroimaging and machine learning communities. Here, we introduce PRoNTo by presenting examples of possible research questions that can be addressed with the machine learning framework implemented in PRoNTo, and cannot be easily investigated with mass univariate statistical analysis. Computer Science Department, University College London ; Institute of Psychology, King's College, London ; Wellcome Trust, London ; Centre de Recherches du Cyclotron - CRC ; Ecole Polytechnique Fédérale de Lausanne Fonds de la Recherche Scientifique (Communauté française de Belgique) - F.R.S.-FNRS ; Fonds pour la formation à la Recherche dans l'Industrie et dans l'Agriculture (Communauté française de Belgique) - FRIA ; PASCAL2 and its Harvest Project ; Department of Computer Science, University College London ; Wellcome Trust ; Portuguese Foundation for Science and Technology , Ministry of Science, Portugal ; Swiss National Science Foundation (PP00P2-123438) and Center for Biomedical Imaging (CIBM) of the EPFL and Universities and Hospitals of Lausanne and Geneva ; The King's College London Centre of Excellence in Medical Engineering, funded by the Wellcome Trust and EPSRC under grant no. WT088641/Z/09/Z Researchers ; Professionals ; Students _URL_ 10.1007/s12021-_PHONE_-1 _URL_ File(s) associated to this reference Fulltext file(s): FileCommentaryVersionSizeAccess Open access Schrouff_Rosa.pdfPublisher postprint1.08 MBView/Open Bookmark and Share SFX Query All documents in ORBi are protected by a user license.
1.217726
EssentialAI/eai-taxonomy-stem-w-dclm-100b-sample
The technique of decorating glass in high relief by cutting or carving away part of the glass between the body of an object and its decoration (e.g., on a cage cup). (from Latin), pl. unguentaria. A term commonly applied to ancient Roman toilet bottles. It appears, however, that the term was "invented" in the 19th century, on analogy with unguentarius (perfume seller) and similar Latin words that the Romans used in connection with perfumes. Glass colored with uranium oxide. The earliest reference to uranium glass appears to date from 1817. It was made popular by Josef Riedel in and after 1834. See also Annagelb (which is yellow) and Annagrün (which is green).
2.363941
HuggingFaceFW/fineweb-edu
As a teacher, you want to engage students and inspire learning. Since teaching has existed, teachers have looked for ways to inspire children to want to learn. Teachers want to take advantage of all resources to equip students to reach their learning potential. Developing a website classroom is a step in the right direction in today's world. Everything is online and adding this layer of portal learning that your students can relate to and possibly embrace this style of learning as it relates to the current relationship students have with computer devices. Knowing what you want to achieve with your classroom website is an important question you should have answered before you begin building your site. This curriculum will help you bring resources and tools that express your lesson plan so teachers, parents, and students know what to expect from learning online. It is another tool that should exist in today's classrooms as we live and interact more with technology. Check out other teachers' websites to get ideas and learn what was successful for them and what were epic failures. Make a point to consider the following as you build your web classroom: Make learning always available Your classroom website should definitely have resources available so students can learn at any time or place. Students have access to the web classroom to gain access to class lessons, worksheets, class notes, assignments and more. Learning does not have to stop because school is over. Students don't have to fail because they missed something during class time. With a web classroom resources of lesson plans are now available for students and their parents while they are at home. This availability of curriculum could change and help that struggling student gain access to class resources at home with the support of their parents. Having access to the web classroom provides extra support with the potential to increase classroom assignments being turned in on time. This could help curb low assignment turn in. Put it all online Use the web classroom to put your entire school year online. Once your basic frame for your site is up you can add your units of study as the year progresses. This way you are not overwhelmed with getting information on your website all at one time. Taking the time to convert your worksheets can be time-consuming, but a worthwhile step to help students gain access to such resources. This also helps you be more effective reaching students in the class. Students become more receptive to teachers who show signs of wanting the student to excel by providing tools to help succeed. By the end of the school year, you have created a complete library of learning tools that other teachers can use and learn from when developing a web classroom for themselves. Web classrooms can offer links to other sites that teach the subject you are teaching in your class. Content and teaching approaches different from yours but effective adds variety to learning for the student. There are puzzles, online quizzes, games and more to connect to your web classroom. These tools are great for completing work assignments and projects in the classroom and assignments at home. Make a point to check out all links before adding them to your web classroom. You want learning tools that enhance what you are teaching. This could help students retain information and hopefully enjoy learning a bit more. This type of information could make a big difference in today's style of learning. Teachers can put things like: – Weekly homework assignments – Upcoming test dates – Field trips – Project assignments This type of service helps the parents as well. They now have a breakdown of what is expected of the student. The parent can play a greater role in the success of the student by having such information made available to them. As the teacher, you could go into detail on certain aspects of the assignments to yield greater understanding. Web classrooms could offer many benefits of continuous learning at home, so students get another opportunity to do well. Publish students work Students that take advantage of this exceptional resource teachers could create a page to showcase students' work where other students could see it. You would need permission first from parent and student. Having a post type page where students' work is on display could inspire higher learning for the student. It could also spark learning for another student. Teacher and parent corner It is a good idea to have a page where co-workers can gain access. Here you can share and exchange ideas to make each web classroom stronger and most effective. A page for parents is often helpful as well. Parents can check in and leave messages or get feedback from teachers on ways to help their child excel in school. A major attraction for web classrooms is the ability to share with others around the world. The Web 2.0 tool named Glogster is a must have for your virtual classroom This software is engaging students to create virtual posters that are vivid, animated and life like. This software inspires the creativity of students and encourages them to combine text, sound, video, images and hyperlinks and share their masterpiece electronically with others. Facebook for web classrooms should include Edmodo as the social network for the classroom. Unlike Facebook, it provides specific restrictions and guides to ensure resource learning is the only objective. Edmodo allows students, teachers and parents to connect and express problems, tips, and ideas. It is here to connect you with the proper source to get solutions that remove learning barriers so students can succeed. The connection of resources parent and teachers may need are available as well so they too can help their student. Private chats between parents, students and teachers are instantly made available. This type of instant support can make all the difference in learning. Learning is all about thinking. It is the goal of every teacher. Bubbl.us is a brainstorming software that helps students get their ideas to grow. One thought from the student is expanded into several common possibilities. The student now has many ideal solutions for their assignments. These ideas are shared with others electronically, so a window for feedback exists. This boosts learning to the next level because students now engage with each other as it relates to their topic of study. Resources for Teachers Teachers want to deliver the best web classroom, so they are always searching for the best tools. A key success to your web classroom are resources. With the software, Tackk teachers can create digital assignments. You can blend your lesson material to deliver on a different level, so the student experiences two styles of learning. Another feature with Tackk is the flipped classroom function. This feature delivers assignment information and helps prep students before arriving for class. Students arrive with questions and ready to dive into the lesson plan of the day. There are many great resources available to create web classrooms. Knowing the needs of your students and what you want to achieve are important steps in creating a format beneficial to teachers, students, and parents.
2.55151
openbmb/Ultra-FineWeb
The stars hold the key to our past. Before people were here, before there was a planet, there was a vast Universe. The power and energy that created the stars may be somewhat of a debate between science and spiritualists, but for many there is a clear understanding that our origins come from the stars - and that we are made of magical star dust. For eons mankind has looked to the stars. Ancient civilizations aligned their buildings and temple structures to galactic alignments and celestial events. At times before humans were considered advanced enough to create modern equipment, they somehow knew about planetary transitions and alignments that occurred at regular, albeit sometimes lengthy intervals. Are we seeking clarity about who we are now by looking out into our distant past? Or do we just love bright, shiny things? I've been taking pictures lately of the sun finding profoundly interesting images. I've done this in the past with the moon as well, to equally interesting results. And although some naysayers will comment that the light refraction off the camera lens is a simple explanation, I challenge them to look deeper and think larger. As aura photography has shown us, we can see energy through the a lens. Colors hold vibratory frequency and energy to them. Each color, a different vibratory reality and symbolism. Look for bright, vibrant colors. Look for orbs. Observe where they are and when the occur. What are you thinking about when you see them? What can they be trying to tell you? For me, it is sometime as simple as a confirmation: I am on the right path when I see vibrant colors and orbs. If I am pondering particular questions when they show up, it is a resounding, "yes." Or sometimes, perhaps, it is an energy download, or an attempt to get my attention. Maybe even, a synchronicity of myself at a different time taking a picture of the sun, reflecting back at my current self. Or, a hint of where we are going in our next evolutionary shift. So many possibilities. I've found myself diving into understanding ancient civilizations on earth, and more often that not, finding starry connections with them. More to come on this. For now, look up and take a shot! Let me know what you see.
1.843132
m-a-p/FineFineWeb
Phase diagram of ESA complexes at 100 mM NaCl salt concentration as a function of pentanol and dodecane content relative to the surfactant weight [wt %]. The natural phenomenon of self-assembly — how molecules or other entities gather to become ordered objects or arrays — occurs in many areas of science, from nanomaterials to biology. Resulting from basic, well understood forces, such as electrostatics, self-assembly may allow scientists to manipulate materials at at tiny scales, ultimately yielding great advances in fields such as data storage, pharmaceuticals, and catalysis. In efforts to better understand self-assembly, many researchers are studying self-assembling systems in solution that consist of an electrolyte (any molecule that dissociates into ions in solution) attached to, or "in complex with," a surfactant, a substance that lowers the surface tension between two liquids (or between a liquid and a solid). These liquid systems, called polyelectrolyte-surfactant complexes (PSCs), are good platforms for scientists to study how polymers and small molecules bind and also how they behave in solution. Looking ahead, scientists are targeting PSCs as a route toward creating functional materials with specific jobs, such as nanostructures for drug delivery. But investigating PSC behavior can be difficult because the complexes self-assemble into phases of many different structures, including cubic assemblies of various-sizes, hexagonal assemblies of cylinders, and stacks. Scientists want to find a method that will allow them to "dial in" a preferred phase. In a significant step toward this ability, researchers working at NSLS recently mapped out the full phase diagrams of a new, enhanced type of PSC, determining all the possible phases given different solution concentrations. The group, which consists of researchers from Stony Brook University, NSLS, and the University of Massachusetts Amherst, call their system an electrostatically self-assembled amphiphilic complex, or ESA. "My lab is interested in creating ordered nanoporous materials for applications in filters, catalysis and fuel-cell membranes," said the study's corresponding scientist, Helmut Strey from Stony Brook University. "This published work is the first step towards this goal by employing self-assembly to create porous materials with tunable pore size." In PSC systems, the different phases are dictated mainly by spherical clusters of surfactant molecules, called micelles, which form when the surfactant is in solution. The phases are governed by the micelles' size and curvature, and how "bendy" they are, and thus these are the characteristics the researchers need to control. At NSLS, the group discovered that adding a cosurfactant called pentanol to the mix allows the surfactant micelles to loosen up. By then incorporating large amounts of an oily liquid called dodecane, the researchers caused each phase's unit cell (the building block of the structure) to nearly double in size. This degree of manipulation is a big step forward in the drive to understand self-assembly and use it to create new materials and technologies. Said Strey, "The advantage of self assembly is that to make a material you just have to throw all the ingredients together and the material forms by itself. In the future, we will develop techniques to solidify our structures to create usable nanoporous materials." The scientists used x-rays to study the system in its various concentrations, employing an advanced technique that allows them to study many samples at once. This high throughput is achieved using a well plate scanner, a sample analysis system that combines a well plate — a tool that can hold many samples at once via with an array of small wells — with a motorized stage that can hold up to three well plates at a time and pass them through the x-ray beam. Using this method scientists can study up to 1,000 samples per day. "This development is exciting because of the possibilities beyond this research," said co-author Elaine DiMasi, NSLS physicist and group leader for the Soft Matter Interface beamline at the future National Synchrotron Light Source II. "As beams become smaller and data acquisition becomes faster, potentially every dataset becomes an 'image,' whether we map out a sample in real space, in 'chemical composition' space, like this work, or in processing space, such as by changing temperature or applying mechanical stress. "Once we can render such data in automatic ways, we have a way to 'light up' regions of interest even in very large parameter spaces, and quickly hone in on the properties of interest to our applications." At NSLS-II, the third-generation light source that will succeed NSLS, the extremely bright x-ray beams produced will enable researchers to increase throughput, approaching a rate of one sample per second! This research is described in the August 25, 2011, online edition of Macromolecules. Summary Slide (.pdf)
2.512818
HuggingFaceFW/fineweb-edu