question
stringlengths
3
300
answer
stringlengths
9
2.77k
context
sequencelengths
7
7
is there any solid reason why we view/record videos and pictures in a rectangular way. and why not just a square?
This has been asked in various ways before though I'm pressed to think up good search terms. A simple answer is that our field of vision of wider than tall so it makes sense for video to be the same.
[ "When a surface is at a high angle relative to the camera, the fill area for a texture will not be approximately square. Consider the common case of a floor in a game: the fill area is far wider than it is tall. In this case, none of the square maps are a good fit. The result is blurriness and/or shimmering, depending on how the fit is chosen. Anisotropic filtering corrects this by sampling the texture as a non-square shape. The goal is to sample a texture to match the pixel footprint as projected into texture space, and such a footprint is not always axis aligned to the texture. Further, when dealing with sample theory a pixel is not a little square therefore its footprint would not be a projected square. Footprint assembly in texture space samples some approximation of the computed function of a projected pixel in texture space but the details are often approximate, highly proprietary and steeped in opinions about sample theory. Conceptually though the goal is to sample a more correct anisotropic sample of appropriate orientation to avoid the conflict between aliasing on one axis vs. blurring on the other when projected size differs.\n", "Square displays are rarely used in devices and monitors. Nonetheless, video consumption on social apps has grown rapidly and led to the emergence of new video formats more suited to mobile devices that can be held in horizontal and vertical orientations. In that sense, square video was popularized by mobile apps such as Instagram and has since been supported by other major social platforms including Facebook and Twitter. It can fill nearly twice as much screen space compared to 16:9 format (when the device is held differently while viewing from how video was recorded).\n", "The impact these films had on the big screen cannot be assessed from television or video, or even from 'scope prints, which marry the three images together with the seams clearly visible. Because they were designed to be seen on a curved screen, the geometry looks distorted on television; someone walking from left to right appears to approach the camera at an angle, move away at an angle, and then repeat the process on the other side of the screen.\n", "Images are usually located from left to right with oldest (in terms of story) on left and newest on right side. For vertical triplet, older image is usually on the top and more recent on bottom. When photos are not related in time, then they are arranged to be most pleasant to viewer's eye (such as dark to bright) or using some logical arrangements.\n", "Because they are taken from an angle, the pixels associated with Pictometry images are trapezoidal, rather than rectangular. This necessitates special software and algorithms to accurately determine objects’ size and position on the maps.\n", "Videos in the wide \"strip\" format can be arbitrarily wide, possibly too wide to fit on a given display. One way to view them is by panning the image horizontally (display only part of the image, panning the frame sideways) while looping through the video. Geometrically this corresponds to replicating the output formula_8 array in the third variable, then cutting diagonally. If the frame advances 1 pixel at a time and the output resolution and frame rate are the same as the input, then the entire strip video will be seen and (with technical assumptions on start and end) the output video will be the same resolution and duration as the input. Rather than looping, which introduces a jump, the video can bounce back and forth between endpoints.\n", "Unfortunately, not all standard TV pictures are exactly 4:3: As mentioned earlier, in analog video, the center of a picture is well-defined but the edges of the picture are not standardized. As a result, some analog devices (mostly PAL devices but also some NTSC devices) generated motion pictures that were horizontally (slightly) wider. This also proportionately applies to anamorphic widescreen (16:9) pictures. Therefore, to maintain a safe margin of error, ITU-R BT.601 required sampling 16 more non-square pixels per line (8 more at each edge) to ensure saving all video data near the margins.\n" ]
Are all the particles in the universe quantum entangled or does that only happen under specific conditions?
I would argue yes as entanglement as well as superposition are intrinsic aspects of a quantum world, but saying that 'everything is entangled' is not a particularly useful or helpful notion. The reason for this is that entanglement happens too much and too easily killing all the delicate quantum effects we so carefully try to cultivate in the laboratory. This process is called decoherence and is a major reason why the weird quantum world generates "classical" physics as we move to larger systems. Here's an excellent enthusiast level look at it: * Zurek. Decoherence and the transition from quantum to classical -- REVISITED. PHYSICS TODAY, 44:36-44 (1991/2003) _URL_0_ What people call EPR entanglement is more correctly described as isolated entanglement. Systems which are allowed to coalesce with each other, but not with the outside world. This only happens under specific conditions.
[ "Quantum mechanics proposes that pairs of virtual particles are being created from quantum fluctuations in this \"empty\" space all the time. If these pairs do not mutually annihilate right away, they could be detected as real particles, for example if one falls into a black hole and its opposite is emitted as Hawking radiation.\n", "It is also possible to create entanglement between quantum systems that never directly interacted, through the use of entanglement swapping. Two independently-prepared, identical particles may also be entangled if their wave functions merely spatially overlap, at least partially.\n", "Quantum mechanics is non-local in the sense that distant systems can be entangled. Entangled states lead to correlations in the results of otherwise random measurements, even when the measurements are made nearly simultaneously and at far distant points. The impossibility of superluminal communication led Einstein, Podolsky, and Rosen to propose that quantum mechanics must be incomplete (see EPR paradox).\n", "Quantum entanglement is a physical phenomenon often invoked for quantum mind models. This effect occurs when pairs or groups of particles interact so that the quantum state of each particle cannot be described independently of the other(s), even when the particles are separated by a large distance. Instead, a quantum state has to be described for the whole system. Measurements of physical properties such as position, momentum, spin, and polarization, performed on entangled particles are found to be correlated. If one of the particles is measured, the same property of the other particle immediately adjusts to maintain the conservation of the physical phenomenon. According to the formalism of quantum theory, the effect of measurement happens instantly, no matter how far apart the particles are. It is not possible to use this effect to transmit classical information at faster-than-light speeds (see Faster-than-light § Quantum mechanics). Entanglement is broken when the entangled particles decohere through interaction with the environment; for example, when a measurement is made or the particles undergo random collisions or interactions. According to David Pearce, \"In neuronal networks, ion-ion scattering, ion-water collisions, and long-range Coulomb interactions from nearby ions all contribute to rapid decoherence times; but thermally-induced decoherence is even harder experimentally to control than collisional decoherence.\" He anticipated that quantum effects would have to be measured in femtoseconds, a trillion times faster than the rate at which neurons function (milliseconds).\n", "Two quantum entangled particles cannot be understood independently. Two or more states in quantum superposition, e.g., as in Schrödinger's cat being simultaneously dead and alive, is mathematically not the same as assuming the cat is in an individual alive state with 50% probability. The Heisenberg's uncertainty principle says that complementary variables, such as position and momentum, cannot both be precisely known – in some sense, they are not individual variables. A \"natural criterion of individuality\" has been suggested.\n", "Quantum systems can become entangled through various types of interactions. For some ways in which entanglement may be achieved for experimental purposes, see the section below on methods. Entanglement is broken when the entangled particles decohere through interaction with the environment; for example, when a measurement is made.\n", "The theory of quantum entanglement predicts that separated particles can briefly share common properties and respond to certain types of measurement as if they were a single particle. In particular, a measurement on one particle in one place can alter the probability distribution for the outcomes of a measurement on the other particle at a different location. If a measurement setting in one location instantaneously modifies the probability distribution that applies at a distant location, then local hidden variables are ruled out. For an expanded description, see Bell's theorem.\n" ]
how does the weather change so drastically for just a few days and then go back?
If you have ever listened to the weather man you've probably heard the term "Cold front" or "Warm front" as they tell you the weather. These fronts are masses of air with various attributes, they could be cold & moist, warm & dry, warm & moist, etc. The leading edges of these fronts are pretty strong and they can push away or go under the current mass of air in an area, when this happens the weather changes. Sometimes these changes are just drops or increases in temperature, sometimes these changes result in lots of steady rain, and sometimes they result in violent thunderstorms. **But the ELI5 is that there are masses of air types floating around and they push and interact with each other and help create our weather.** So right now a cold front has entered your area and probably has gone under the warm air you've been having and once that cold air has left your warm air will will return (although it really never left, in some cases it was just above the cold front!)
[ "Typically, if there is little wind shear, the storm will rapidly enter the dissipating stage and 'rain itself out', but, if there is sufficient change in wind speed or direction, the downdraft will be separated from the updraft, and the storm may become a supercell, where the mature stage can sustain itself for several hours.\n", "The main influence on U.S. weather is the polar jet stream which migrates northward into Canada in the summer months, and then southward into the US in the winter months. The jet stream brings in large low pressure systems from the northern Pacific Ocean that enter the US mainland over the Pacific Northwest. The Cascade Range, Sierra Nevada, and Rocky Mountains pick up most of the moisture from these systems as they move eastward. Greatly diminished by the time they reach the High Plains, much of the moisture has been sapped by the orographic effect as it is forced over several mountain ranges.\n", "Weather is a chaotic system that is readily modified by small changes to the environment, so accurate weather forecasting is limited to only a few days. Overall, two things are happening worldwide: (1) temperature is increasing on the average; and (2) regional climates have been undergoing noticeable changes.\n", "During the last week of March, meteorologists expected average temperatures to return sometime from April to mid-May. On April 10, 2014, a ridge of high pressure moved into the Eastern United States, bringing average and above-average temperatures to the region, which ended the cold wave.\n", "Within 36 hours of the passage of a tropical cyclone's center, the pressure begins to fall and a veil of white cirrus clouds approaches from the cyclone's direction. Within 24 hours of the closest approach to the center, low clouds begin to move in, also known as the bar of a tropical cyclone, as the barometric pressure begins to fall more rapidly and the winds begin to increase. Within 18 hours of the center's approach, squally weather is common, with sudden increases in wind accompanied by rain showers or thunderstorms. Within six hours of the center's arrival, rain becomes continuous. Within an hour of the center, the rain becomes very heavy and the highest winds within the tropical cyclone are experienced. When the center arrives with a strong tropical cyclone, weather conditions improve and the sun becomes visible as the eye moves overhead. Once the system departs, winds reverse and, along with the rain, suddenly increase. One day after the center's passage, the low overcast is replaced with a higher overcast, and the rain becomes intermittent. By 36 hours after the center's passage, the high overcast breaks and the pressure begins to level off.\n", "Starting in late December, a change in atmospheric pressure shifts ocean currents so that warm waters come closer to shore and displace the cold waters. During this time, air and water temperatures, tides, sea levels and wave heights, and relative humidity all rise. These conditions produce heavy rainfall that used to last through August, but now usually only lasts into May. Rain during the rainy season is punctuated by sunny weather. \n", "There are new transformative weather effects in the expansion, including wind. The sky is an indicator of what weather will take place; for example, clouds will get darker, which means the storm is approaching. Gardening can be affected by the weather. Snow and rain will gather on roadways. Cars will not slip off the road. Temperature will control how easily snow will accumulate on the ground or trees, or how quickly waters such as puddles will evaporate. In the Fall Sims can rake leaves.\n" ]
why is being transgender not considered sexist?
It isn't about the things you like to do, it's about who you feel you are. These things are quite different. Like you said, if a boy wants to play with barbies or a girl wants a toy gun, whatever! It doesn't have anything to do with being transgender. I think the confusion for you might be coming from the fact that if children exhibit these traits and then grow up to be transgender it's not because those things are inherently "girl things" or "boy things" but because culture often dictates that they are. So if a little boy feels like a girl (but doesn't know how to express it) he will do things that in his mind feel right, and those things will often be things that his culture has deemed "girl things".
[ "Many transgender people see gender identity as an innate part of a person. Some feminists also criticize this belief, arguing instead that gender roles are societal constructs, and are not related to any natural factor. Sex-positive feminists support the right of all individuals to determine their own gender and promote gender fluidity as one means for achieving gender equality. Patrick Califia has written extensively about issues surrounding feminism and transgender issues, especially in \"Sex Changes: Transgender Politics\".\n", "BULLET::::- Acts of Violence, Discrimination, and Other Abuses Based on Sexual Orientation and Gender Identity\"Consensual same-sex sexual activity is illegal under the law. LGBTI groups reported police used the law as a pretext to bully LGBTI individuals, as well as those considered effeminate regardless of their sexual orientation, and to limit registration of LGBTI organizations. Some groups also reported harassment under a suspicious behavior provision of the police code. The transgender population has long been a marginalized, but recognized part of society, but it faced continued high levels of fear, harassment, and law enforcement contact in the wake of violent extremist attacks against vulnerable communities. Members of LGBTI communities received threatening messages via telephone, text, and social media, and some were harassed by police. In May, RAB forces raided the Chayaneer Community Center in Keranigan during a dinner organized by the LGBTI community from that area. According to witnesses, 28 individuals were arrested of the 120 persons present at the time of the raid. The witnesses also stated RAB separated the diners into small groups and beat them before identifying individuals for arrest. During the raid RAB announced to the media the raid was conducted based on suspicion of homosexual activity and allowed the media to photograph some of the arrested individuals. RAB later announced the attendees were not engaged in “illegal sexual activities” at the time of the raid and were instead arrested for possession of narcotics—specifically \"yaba\" (a combination of methamphetamine and caffeine) and cannabis. The court system remanded four of the individuals. Of the remaining 24 individuals, 12 were detained for further questioning and 12 were sent directly to jail. Following these events and continued harassment, many members of LGBTI communities, including the leadership of key support organizations, continued to reduce their activities and sought refuge both inside and outside of the country. This resulted in severely weakened advocacy and support networks for LGBTI persons. Organizations specifically assisting lesbians continued to be rare. Strong social stigma based on sexual orientation was common and prevented open discussion of the subject.\"\n", "Feminist views on transgender people differ. Some feminists do not view trans women as women, believing that they have male privilege due to their sex assignment at birth. Additionally, some feminists reject the concept of transgender identity due to views that all behavioral differences between genders are a result of socialization. In contrast, other feminists and transfeminists believe that the liberation of trans women is a necessary part of feminist goals. Third-wave feminists are overall more supportive of trans rights. A key concept in transfeminism is of transmisogyny, which is the irrational fear of, aversion to, or discrimination against transgender women or feminine gender-nonconforming people.\n", "Other transgender rights authors argue that a significant part of the oppositional sexist origin of transphobia, and especially of the forms that incite violence towards transsexual people, is linked to psychological claims of difference between male sexuality and female sexuality in the brain's protection mechanisms from committing sex crimes. These authors argue that the assumption that men's acceptable sexuality is based on category-specific sexual arousal while women's acceptable sexual behavior is said to be due to lower sex drive and especially higher sexual inhibitions causes allegations that transsexual people have neither safety system in the brain and are sex criminals, and recommend information about flaws in studies that claim to show such sex differences (including the possibility that fear of being alleged to be inappropriately sexually aroused may deter more men than women from taking part in sexual arousal studies) as a remedy.\n", "A common misconception is that a transgender person is therefore gay. However, being transgender focuses on gender identity and not sexual orientation. A transgender person may identify with any sexual orientation. Another important misconception is that individuals who crossdress are transgender. However, many crossdressers are comfortable with their assigned sex. Even though individuals who participate in crossdressing are officially under the Transgender Umbrella, most do not identify as transgender.\n", "In Chile, transgender people are often associated with homosexuality. Transgender women are mostly discriminated, unable to enter the labor market so their only way of survival is prostitution, therefore they are exposed to violence and police harassment. According to a 2009 study conducted by RedLacTrans, about 95% of Chilean transgender women work as prostitutes because of family, social and employment discrimination. As for transgender men, in many cases they decide to violate their gender identity using female or unisex clothes to get a job. Since 2007, name and sex change on legal documents are only allowed by court order.\n", "Some transgender and transsexual people dislike the term \"sexual minority\" for yet another reason. They argue that the phenomenon of transsexuality or transgender has nothing to do with sex, sexual practices or sexual orientation, but it relates to the gender, gender dysphoria and gender-variant behavior or feelings. Thus, they feel it is incorrect to classify them as \"sexual minority\", when, in fact, they are gender-variant minority.\n" ]
If forced to spin, whether for a game or as a contest, what is the most effective way to avoid sever dizziness?
Do what ballerinas do. Keep your head in one position and turn it fast and stop, fast and stop, fast and stop. This is precisely why they do this - to not get dizzy.
[ "In some trick games--typically ones in which players are not penalized for winning tricks, and there is no requirement for trumping or following suit when possible--players may \"slough\", or play a card face down. A card so played is incapable of winning the trick; but sloughing has the advantage that the other players cannot see what card is played. As this form of sloughing has the potential to be used to cheat in most games (i.e. playing a winning card face-down to avoid taking an \"overtrick\" or a trick containing penalty points) and is thus not allowed, \"sloughing\" in the vernacular more often refers to simply discarding an off-suit card on a trick, particularly one that could be dangerous to that player if kept. This form of sloughing is important in evasion games and in some contract games where \"overtricks\" are penalized; in Oh Hell, for instance, a player who cannot follow suit may elect to discard a card that would win if played to follow suit later, thus reducing the chance that the player will \"bag\", or take more tricks than needed. This is common in Hearts, where high-value cards (especially Spades and Hearts) are dangerous as they increase the chance of winning a trick with penalty points.\n", "The act of putting an opponent on tilt may not pay off in the short run, but if some time is put into practicing it, a player can quickly become an expert at “tilting” other players (with or without using bad manners). In theory, the long-run payoff of this tactic is a monetarily positive expectation.\n", "The act of deliberately misplaying a hand in order to induce a mistake by an opponent which results in either the same or a superior result. Even when the gambit does not yield a material gain, it usually induces a big psychological impact on the opponents who were offered a trick for free but couldn't have believed it were possible.\n", "Players can try to hold on to their trick cards until needed because if they are used early, the opponent can tell if the player has any trick cards to block any moves. This can be risky, however, as most versions of the rules state that the game cannot end on a trick card. \n", "It must be done by rotating the head as close to perfectly in the horizontal plane as possible so as not to defeat the purpose of minimizing dizziness in those so predisposed. The most common spotting is 180° to and away from one's partner, or the LOD and a full 360° from the original spot, be it LOD, OLOD, or toward or away from one's partner, a wall for example. It can be done in apart/free position or less frequently in Closed Position.\n", "One way to mask the symptoms of the spins is to avoid staring at moving objects, such as people who are dancing or ceiling fans. Instead, it helps to stare at a non-moving object and slowly blink a few times. However, it will make things worse to keep one's eyes closed for an extended period. In minor cases of the spins, simply sitting alone in a quiet place or taking a walk is all it takes to make them subside.\n", "In tricking, this move can be done from a run, a step in, a standing position, or from another move in a combo. A standing position and linking it in or at the end of a combo are more difficult due to the height and amount of torque needed. In Capoeira, it can be done from a ginga or from another move in a combo.\n" ]
Wondering about the origins of current Japanese cultural norms
Japanese people show emotion... they're just more subdued about it, and you need to look for non-verbal clues a bit more than you're probably used to. It was considered a loss of face to show strong emotions in public, if you're asking about the origin of this.
[ "Japanese culture has evolved greatly from its origins. Contemporary culture combines influences from Asia, Europe and North America. Traditional Japanese arts include crafts such as ceramics, textiles, lacquerware, swords and dolls; performances of bunraku, kabuki, noh, dance, and rakugo; and other practices, the tea ceremony, ikebana, martial arts, calligraphy, origami, onsen, Geisha and games. Japan has a developed system for the protection and promotion of both tangible and intangible Cultural Properties and National Treasures. Twenty-two sites have been inscribed on the UNESCO World Heritage List, eighteen of which are of cultural significance.\n", "The next major period in Japanese history was the Muromachi Period, pointing to the rise of , centered around the gorgeous cultural world of Ashikaga Yoshimitsu and his villa in the northern hills of Kyoto (Kinkaku-ji), and later during this period, the rise of Higashiyama Culture, centered around the elegant cultural world of Ashikaga Yoshimasa and his retirement villa in the eastern hills of Kyoto (Ginkaku-ji). This period, approximately 1336 to 1573, saw the budding of what is generally regarded as Japanese traditional culture as we know it today.\n", "Japanese culture begins to develop in no small part due to influences from mainland trade and immigration from China. During this time in the pre-writing historical period, objects from the mainland start appearing in large amounts, specifically mirrors, swords, and jewels. All three of these have a direct connection to the imperial divine status as they are the symbols of imperial divinity and are Shinto honorary objects. Also the rice culture begins to blossom throughout Japan and this leads to the settlement of society, and seasonal reliance of crops. Both of these changes are highly influential on the Japanese people's relationship to the natural world, and likely development of a more complex system of religion. This is also the period that is referenced as the beginning of the divine imperial family. The Yayoi culture was a clan based culture that lived in compounds with a defined leader who was the chief and head priest. They were responsible for the relationship with their \"gods\" Kami and if one clan conquered another, their \"god\" would be assimilated. The earliest records of Japanese culture were written by Chinese traders who described this land as \"Wa\". This time period led to the creation of the Yamato culture and development of formal Shinto practices.\n", "The inhabitants of Japan experienced a long period of relative isolation from the outside world during the Tokugawa shogunate after Japanese missions to Imperial China, until the arrival of the \"Black Ships\" and the Meiji period. Today, the culture of Japan stands as one of the leading and most prominent cultures around the world, mainly due to the global reach of its popular culture.\n", "During this time, Japan remained isolated from the world, so its culture developed with very little foreign influence. One of the major cultural movements of the Tokugawa period was the institution of a branch of scholarship called \"kokugaku\" (), literally \"National Studies\" and commonly translated as \"Japanese Studies\". Practitioners of the movement, or \"Kokugakushu\", attempted to distinguish between what was genuine Japanese culture and what was foreign culture, and to restore Japanese culture to what it was before the influence of foreigners—especially the Chinese. Their work had a large focus on Shinto, Japan's indigenous religion. Early Tokugawa Confucians tried to link Shinto with China by pinpointing its Chinese origins. The Hirata school of the \"kokugaku\" movement responded by initiating a project to \"Japanize\" the \"I Ching\", a book that was a major influence on Shinto, by claiming it was of Japanese origin. The project was completed with Aizawa Seishisai emptying the \"I Ching\" of its Chinese content. The rise in national self-respect during this time resulted in Japan viewing itself as the center of a \"civilized world surrounded by barbarians.\"\n", "Since the Meiji Restoration, Japan has absorbed elements of Western culture and has given them a \"Japanese\" feel or modification into it. Its modern decorative, practical and performing arts works span a spectrum ranging from the traditions of Japan to purely Western modes. Products of popular culture, including J-pop, J-rock, manga and anime have found audiences and fans around the world.\n", "Japanese popular culture not only reflects the attitudes and concerns of the present day, but also provides a link to the past. Popular films, television programs, manga, music, anime and video games all developed from older artistic and literary traditions, and many of their themes and styles of presentation can be traced to traditional art forms. Contemporary forms of popular culture, much like the traditional forms, provide not only entertainment but also an escape for the contemporary Japanese from the problems of an industrial world.\n" ]
Looking to learn more about Canadian History
From my perspective — and it's biased toward the Yukon/Alaska — I'd recommend working your way through Pierre Berton's bibliography. While he wrote some duds here and there, most of his 50 titles are pretty good. I'd recommend *The Klondike Fever*, *The Promised Land: Settling the West* and *Marching As To War* as three to get you started. Those three condense a lot of his bibliography. Once you've gotten through those, grab his two volumes on the War of 1812 (one covers 1812-1813, the other 1814-1815) his two volumes on the Canadian Pacific Railway, *The Arctic Grail*, *Vimy*, *Niagra* and *The Great Depression.*
[ "BULLET::::- Schultz, John. ed. \"Writing About Canada: A Handbook for Modern Canadian History\" (1990), chapters by experts on politics, economics, ideas, regions, agriculture, business, labor, women, ethnicity and war.\n", "The historiography of Canada deals with the manner in which historians have depicted, analyzed, and debated the history of Canada. It also covers the popular memory of critical historical events, ideas and leaders, as well as the depiction of those events in museums, monuments, reenactments, pageants and historic sites.\n", "Historica Canada is the country's largest organization dedicated to enhancing awareness of Canadian history and citizenship. All of its programs are offered bilingually and reached more than 28 million Canadians last year. \n", "\"David McCord created one of the earliest and most important collections of objects, images and manuscripts associated with the history of Canada. Convinced that an understanding of the past strengthens national identity, he devoted most of his life and personal fortune to gathering and documenting some 15,000 items related to Aboriginal, French and British history in North America. This man of vision bequeathed his outstanding collection to McGill University, which fulfilled his dream of founding a museum for the benefit of all Canadians.\"\n", "BULLET::::- \"Canada: A People's History\", Volume 1 and 2 (in collaboration with Don Gillmor), Ed. McClelland&Steward et Editions Fidès 2000. Winner of the Ex-Libris Price of The Association of Canada's Bookstores. Mention of the best book on Canada's history given by the ministry of \"PCanadian Heritage\".\n", "The Canadian Historical Association (CHA; French \"Société historique du Canada\", SHC) is a Canadian organization founded in 1922 for the purposes of promoting historical research and scholarship. It publishes the \"Journal of the Canadian Historical Association\", \"The CHA Bulletin\" and a well-respected series of booklets featuring concise treatments of particular aspects of Canadian history.\n", "BULLET::::- 1959. A bibliography of Canadiana : being items in the Public Library of Toronto, Canada, relating to the early history and development of Canada. Edited by Gertrude M. Boyle, assisted by Marjorie Colbeck ; with an introduction by Henry C. Campbell.\n" ]
why as seen on tv products are sold in "3 payments of $19.95" or something like that.
Primarily for two reasons: 1) Splitting up the cost of a product over multiple payments gives the potential buyer the impression that the product is actually cheaper than it is. Ex. One payment of $60 vs three payments of $19.99. The fact that the sellers do not round up is also by intention. Even though we see $19.99 or $19.95 we don't equate that with $20, our immediate association puts it with $19 which seems cheaper to us. 2) There is some flexibility with these payments as by being able to pay in installments reduces a small portion of the financial burden on the buyer - making him/her more likely to buy it. It's like saying "Well, okay, I can pay $19 bucks this month and get the product, and worry about the two other payments later!" Out of sight - out of mind. Hope this helps!
[ "The product is sold direct to consumers in these countries using an instalment-based payment model. This ensures that the product is affordable to people who would be unable to afford a large lump sum payment.\n", "BULLET::::- Product A currently sells for $10. The seller decides to increase the price to $15, but finds that he ends up making less money. This is because he is selling fewer of the product due to the increased price, and his total revenue has fallen. The demand for this product must be elastic.\n", "Few consumers are aware of the fact that they are the ones paying for every cent spent for public relations, advertisements, rebates, packaging etc., since they ordinarily get included in the price calculation.\n", "At launch on December 2, 2007, set-top boxes were available for prices ranging between R$900 (~US$450) and R$1200 (~US$600), inhibiting sales. But after 8 months the prices dropped quickly to around R$300 (~US$150). The Federal Government announced subsidies worth 1 billion Reais (~US$556 millions) so these prices will face a new reduction phase.\n", "The sets were released during the dot-com bubble and a healthy economic environment in the United States. The figures themselves cost between $5.99 USD and $7.99 USD each, while the playsets generally ranged from $20 USD to $25 USD at the end of 2004. There were also special releases that were in the $60 USD range. Even at minimal retail, 200+ figures and 40 playsets meant a large financial commitment. In addition, there were a dozen exclusives and mail-in figures that required extra work to obtain the figures — often via an online auction. However, the early success of the series and the healthy economy added extra momentum to the line for a few more years.\n", "BULLET::::- Additional payments (on top of the main price of a purchase) which would need to have active or express consent of the consumer. An example is that pre-ticked boxes which the consumer must 'untick' will no longer be allowed\n", "The US released figures are usually valued from $10US to $20US in great condition. The UK line, especially the UK exclusives are considerably more valuable, with the loose figures ranging from $5US to $30US and MOC figures from $25US to more than $200US\n" ]
Were wolves domesticated only once, with all dogs descended from that single progenitor population, or did the process of domesticating wolves occur independently in different societies?
While we have a fair bit of overlap, given that this predates written history, you may want to consider X-posting this to /r/AskAnthropology as well as this falls within their purview. And for that matter, I would note that /r/AskScience would also be an appropriate venue given the biological science aspects of the question.
[ "An examination of the genetic evidence by Carles Vila and others confirms that the progenitor of the domestic dog is the wolf (\"Canis lupus\"). The suggested date of their domestication is about 100,000 BC. While accepting the wolf as the ancestor, paleontologists and archaeologists believe domestication came much later. A reconstruction by a joint team of researchers from China and Sweden postulates that humans may have domesticated dogs from wolves as recently as 15,000 years ago. They found that, while most dogs share a common gene pool, genetic diversity is highest in East Asia, suggesting dogs have been domesticated there the longest. Genetic research suggests that all domestic dogs worldwide may have originated from possibly three female wolves.\n", "The hypothesis is that two genetically different, and possibly now extinct, wolf populations were domesticated independently in eastern and western Eurasia to produce paleolithic dogs. The eastern Eurasian dogs then dispersed westward alongside humans, reaching western Europe 6,400–14,000 years ago where they partially replaced the western paleolithic dogs. A single domestication is thought to be due to chance, however dual domestication on different sides of the world is unlikely to have happened randomly and it suggests that external factors - an environmental driver - may have forced wolves to work together with humans for survival. It is possible that wolves took advantage of resources that humans had, or humans may have been introduced to wolves in an area in which they didn't previously live.\n", "The genetic divergence between dogs and wolves occurred between 40,000–20,000 years ago, just before or during the Last Glacial Maximum. This timespan represents the upper time-limit for the commencement of domestication because it is the time of divergence and not the time of domestication, which occurred later. The domestication of animals commenced over 15,000 years ago, beginning with the grey wolf (\"Canis lupus\") by nomadic hunter-gatherers. The archaeological record and genetic analysis show the remains of the Bonn–Oberkassel dog buried beside humans 14,200 years ago to be the first undisputed dog, with disputed remains occurring 36,000 years ago. It was not until 11,000 years ago that people living in the Near East entered into relationships with wild populations of aurochs, boar, sheep, and goats.\n", "The genetic divergence between dogs and wolves occurred between 40,000–20,000 years ago, just before or during the Last Glacial Maximum. This timespan represents the upper time-limit for the commencement of domestication because it is the time of divergence and not the time of domestication, which occurred later. The domestication of animals commenced over 15,000 years ago, beginning with the grey wolf (\"Canis lupus\") by nomadic hunter-gatherers. The archaeological record and genetic analysis show the remains of the Bonn–Oberkassel dog buried beside humans 14,200 years ago to be the first undisputed dog, with disputed remains occurring 36,000 years ago. It was not until 11,000 years ago that people living in the Near East entered into relationships with wild populations of aurochs, boar, sheep, and goats.\n", "Genetic studies suggest a domestication process commencing over 25,000 YBP, in one or several wolf populations in either Europe, the high Arctic, or eastern Asia. The remains of large carcasses left by human hunter-gatherers may have led some wolves into entering a migratory relationship with humans. This could have led to their divergence from those wolves that remained in the one territory. A closer relationship between these wolves — or proto-dogs — and humans may have then developed, such as hunting together and mutual defence from other carnivores and other humans. Around 10,000 YBP agriculture was developed resulting in a sedentary lifestyle, along with phenotype divergence of the dog from its wolf ancestors, including variance in size. In the Victorian era, directed human selection developed the modern dog breeds, which resulted in a vast range of phenotypes. Each of these domestication phases have left their mark on the dog's genome.\n", "While humans may have intentionally domesticated wolves into dogs, an alternate hypothesis is that wolves effectively domesticated themselves by establishing a mutually beneficial relationship with prehistoric humans. They scavenged on the remains of the prey animals left by the prehistoric people at the human settlements or the kill sites. Those wolves that were less anxious and aggressive thrived, continued to follow the prehistoric humans, and colonized the human-dominated environments, generation after generation. Gradually, the first primitive dogs emerged from this group.\n", "There is clear evidence that dogs were derived from gray wolves during the initial phases of domestication and that no other canine species was involved. The wolf population(s) that were involved are likely to be extinct. Despite numerous genetic studies of both modern dogs and ancient dog remains, there is no firm consensus regarding either the timing or location(s) of domestication, the number of wolf populations that were involved, or the long-term effects domestication has had on the dog's genome.\n" ]
Are fields real or just a mathematical tool?
The strongest statement physics can ever make is "this model makes accurate predictions." It cannot tell you whether the model is what's "really" going on or not, because there is no other way to test the model. Whether something is "real" or not is a matter of philosophy. You could argue that any model which makes perfect physical predictions is reality because all correct descriptions of reality are equivalent, sort of like they're just different languages expressing the same concept. I dunno, it certainly isn't science at that point.
[ "When mathematicians employ the field axioms, the intentions are even more abstract. The propositions of field theory do not concern any one particular application; the mathematician now works in complete abstraction. There are many examples of fields; field theory gives correct knowledge about them all.\n", "The best known fields are the field of rational numbers, the field of real numbers and the field of complex numbers. Many other fields, such as fields of rational functions, algebraic function fields, algebraic number fields, and \"p\"-adic fields are commonly used and studied in mathematics, particularly in number theory and algebraic geometry. Most cryptographic protocols rely on finite fields, i.e., fields with finitely many elements.\n", "Fields play an important role in science, technology, and economy. They describe the spatial variations of a quantity, like the air temperature, as a function of position. Knowing the configuration of a field can be of large value. Measurements of fields, however, can never provide the precise field configuration with certainty. Physical fields have an infinite number of degrees of freedom, but the data generated by any measurement device is always finite, providing only a finite number of constraints on the field. Thus, an unambiguous deduction of such a field from measurement data alone is impossible and only probabilistic inference remains as a means to make statements about the field. Fortunately, physical fields exhibit correlations and often follow known physical laws. Such information is best fused into the field inference in order to overcome the mismatch of field degrees of freedom to measurement points. To handle this, an information theory for fields is needed, and that is what information field theory is.\n", "It is possible to construct simple fields without any prior knowledge of physics using only mathematics from several variable calculus, potential theory and partial differential equations (PDEs). For example, scalar PDEs might consider quantities such as amplitude, density and pressure fields for the wave equation and fluid dynamics; temperature/concentration fields for the heat/diffusion equations. Outside of physics proper (e.g., radiometry and computer graphics), there are even light fields. All these previous examples are scalar fields. Similarly for vectors, there are vector PDEs for displacement, velocity and vorticity fields in (applied mathematical) fluid dynamics, but vector calculus may now be needed in addition, being calculus over vector fields (as are these three quantities, and those for vector PDEs in general). More generally problems in continuum mechanics may involve for example, directional elasticity (from which comes the term \"tensor\", derived from the Latin word for stretch), complex fluid flows or anisotropic diffusion, which are framed as matrix-tensor PDEs, and then require matrices or tensor fields, hence matrix or tensor calculus. The scalars (and hence the vectors, matrices and tensors) can be real or complex as both are fields in the abstract-algebraic/ring-theoretic sense.\n", "Fields serve as foundational notions in several mathematical domains. This includes different branches of analysis, which are based on fields with additional structure. Basic theorems in analysis hinge on the structural properties of the field of real numbers. Most importantly for algebraic purposes, any field may be used as the scalars for a vector space, which is the standard general context for linear algebra. Number fields, the siblings of the field of rational numbers, are studied in depth in number theory. Function fields can help describe properties of geometric objects.\n", "A field is an algebraic structure in which multiplication, addition, subtraction, and division are all well-defined and satisfy their familiar properties. The real numbers, for example, form a field, as do the complex numbers and rational numbers. Unlike these familiar examples of fields, which are all infinite sets, some fields have only finitely many elements. The simplest example is the field with two elements formula_148 with addition defined by formula_149 and formula_150, and multiplication formula_151 and formula_152.\n", "Many results valid for function fields also hold, at least if reformulated properly, for number fields. However, the study of number fields often poses difficulties and phenomena not encountered in function fields. For example, in function fields, there is no dichotomy into non-archimedean and archimedean places. Nonetheless, function fields often serves as a source of intuition what should be expected in the number field case.\n" ]
why is it more common for employers to provide health insurance plans as opposed to individuals getting their own insurance? why doesn't this group approach differ from other types of insurance, like home, life and auto, where customers often get insurance directly with an insurance company?
This is an accidental remnant of wage controls during the world war. Certain industries could only offer so much salary, ao they started offering healthcare as an artificial raise. Then that became expected. Then it started to become law. Businesses could negotiate good rates, and paying 5k for someone's healthcare saved you the taxes that having to pay that in salary would cost. Eventually you were told that the goal of adulthood was to find a job that offered a good salary and healthcare. It really isnt a system anyone would set up intentionally, but it accidently started and has slowly become too "normal" for most people to consider changing it.
[ "In the article \"Why Tie Health Insurance to a Job?\", Emanuel said that employer based health insurance should be replaced by state or regional insurance exchanges that pool individuals and small groups to pay the same lower prices charged to larger employers. Emanuel said that this would allow portable health insurance even to people that lose their jobs or change jobs, while at the same time preserving the security of employer based health benefits by giving consumers the bargaining power of a large group of patients. According to Emanuel, this would end discrimination by health insurance companies in the form of denial of health insurance based on age or preexisting conditions. In \"Solved!\", Emanuel said that Universal Healthcare Vouchers would solve the problem of rapidly increasing health care costs, which, rising at three times the rate of inflation, would result in higher copayments, fewer benefits, stagnant wages and fewer employers willing to pay for health care benefits.\n", "Group insurance is an insurance that covers a group of people, for example the members of a society or professional association, or the employees of a particular employer for the purpose of taking insurance. Group coverage can help reduce the problem of adverse selection by creating a pool of people eligible to purchase insurance who belong to the group for reasons other than the wish to buy insurance, which might be because they are a worse than average risk. Grouping individuals together allows insurance companies to give lower rates to companies, \"Providing large volume of business to insurance companies gives us greater bargaining power for clients, resulting in cheaper group rates.\"\n", "BULLET::::- Health insurance companies are not actually providing traditional insurance, which involves the pooling of risk, because the vast majority of purchasers actually do face the harms that they are \"insuring\" against. Instead, as Edward Beiser and Jacob Appel have separately argued, health insurers are better thought of as low-risk money managers who pocket the interest on what are really long-term healthcare savings accounts.\n", "Most of the individuals who buy health insurance are related to people covered by employer-sponsored schemes, with the rest accounting for less than 2% of all policies. As companies cut back on their healthcare policies, employees in some cases resort to buying them individually.\n", "Since the majority (85%) of Americans have health insurance, they do not directly pay for medical services. Insurance companies, as payors, negotiate health care pricing with providers on behalf of the insured. Hospitals, doctors, and other medical providers have traditionally disclosed their fee schedules only to insurance companies and other institutional payors, and not to individual patients. Uninsured individuals are expected to pay directly for services, but since they lack access to pricing information, price-based competition may be reduced. The introduction of high-deductible insurance has increased demand for pricing information among consumers. As high-deductible health plans rise across the country, with many individuals having deductibles of $2500 or more, their ability to pay for costly procedures diminishes, and hospitals end up covering the cost of patients care. Many health systems are putting in place price transparency initiatives and payments plans for their patients so that the patients better understand what the estimated cost of their care is, and how they can afford to pay for their care over time.\n", "Can insurance companies maintain profitability in the years to come? Insurance companies still rely on the employer based book of business which is in constant flux with the implementation of the Affordable Care Act, some insurance companies need to reduce premiums due to regulatory pressures on excess revenue, the market place itself is changing from insured to self-insured, from comprehensive to high-deductible plan design.\n", "Individuals would be able to buy their insurance from an exchange created in their state, which would be created in 2010. Premiums would be capped at 13% of the buyer's income, with the rest of the costs paid by subsidies. Insurance companies would be required to cover certain services such as hospitalization, maternity care, newborn care, chemotherapy, and pediatric care. Their plans would be presented as either bronze, silver, gold, and platinum options, sorted by the least to the most expensive. Only individuals and firms with fifty or fewer employees would be eligible to use the exchanges until 2017, when states would bring in larger employers. In 2022, they would become open to all. Persons who are not lawfully in the United States would not be permitted to participate in the exchange, either with or without subsidies. According to Baucus: In order to prevent undocumented immigrants from accessing the state exchanges obtaining federal health care tax credits, the Chairman's Mark requires verification of the following personal data. Name, social security number, and date of birth will be verified with Social Security Administration (SSA) data. For individuals claiming to be U.S. citizens, if the claim of citizenship is consistent with SSA data then the claim will be considered substantiated.\n" ]
How do fields form, especially those right next to densely wooded areas?
Can you give an example of this?
[ "Seemingly treeless fields may contain seeds and living tree stumps and roots which have the ability to sprout new stems and regenerate trees. Even this 'bare' millet field in West Africa contains hundreds of living stumps per hectare which are buried beneath the surface like an underground forest.\n", "In agriculture, a field is an area of land, enclosed or otherwise, used for agricultural purposes such as cultivating crops or as a paddock or other enclosure for livestock. A field may also be an area left to lie fallow or as arable land.\n", "Three main types of soil form bands from north to south. However in the fields to the northeast, i.e. mostly made up of Upper Gatton Park and Farm, there is a round area of \"free-draining, slightly acid, sandy soil\" which as the Soilscape study indicates is used for dry pasture, coniferous woodland and, as not stated there, for wheat. One wood includes a small section in the east the \"freely draining slightly acid loamy soil\" that provides the infill of the Downs and is seen for example in Kingswood and counterintuitively all of the southern developed parts of Reigate, Redhill, most of Dorking, much and all of central Croydon and all around the lowland commons, heaths and parks of south-west London. Throughout the north, the soil is fertile \"slightly acid loamy and clayey soil with impeded drainage\", interrupted by a steep narrow middle course of mid-fertility \"shallow lime-rich soils over chalk or limestone\" which the M25 motorway passes through in Gatton, producing a sheer chalk face on the northern bank east of the Junction 8 interchange, whereas on its very gradual descent to the north-west close to Leatherhead grassed banks surround a far more shallow chalk cutting of the M25 — the only other locations where the M25 passes through chalk slopes are approaching Junction 4 close to Orpington, Halstead and Shoreham, Kent. Across the south of Gatton Park and Reigate Hill Golf Course soil is \"slowly permeable, seasonally wet slightly acid but base-rich, loamy and clayey\".\n", "The Great Field lies on a bend in the river, bounded on three sides by water and to the north by wilderness. It is populated by a growing number of tents, and its story is told by the unnamed narrator - one of the first to pitch camp. As the year advances from Spring to Summer each new arrival adds complexity to the fledgling society, its unwritten rules and petty jealousies. A large regimented group set up a tent village in the south of the field and digs a drainage ditch which divides both the field and opinion and the narrator is forced to take sides. Further groups arrive and threaten the delicate balance of power, then Hippo - clad in just a coarse blanket brings a message that he proclaims all must hear...\n", "A Headland, in agriculture, is the area at each end of a planted field. In some areas of the United States, this area is known as the Turnrow. It is used for turning around with farm implements during field operations and is the first area to be harvested to minimize crop damage. The rows run perpendicular to the lay of the field and are usually two, three or four times the width of the implement used for planting the field.\n", "The water source of the fields comes from the upper streams and waterfalls. When there are low points in the mountain it is important to overcome this by moving water from higher places. Bamboo is cut in half and used as a tool to transfer water using natural gravity into the fields, according to Hmong People’s experience of working on it, the water is moved into the first terrace then a gate is opened to make the water flow into the next terrace using a gate. This process avoids flooding the fields and retains the soil fertility.\n", "The characteristic topography of much old-growth forest consists of pits and mounds. Mounds are caused by decaying fallen trees, and pits (tree throws) by the roots pulled out of the ground when trees fall due to natural causes, including being pushed over by animals. Pits expose humus-poor, mineral-rich soil and often collect moisture and fallen leaves, forming a thick organic layer that is able to nurture certain types of organisms. Mounds provide a place free of leaf inundation and saturation, where other types of organisms thrive.\n" ]
why does it sound odd shortening "you are" to "you're" in the sentence "i'm ready when you are"?
Because in this form the word *you* is emphasized, spoken more loudly. So also emphasizing the *are* leaves the listener wondering: *are* what?
[ "The infinitives of 'I am' are ' 'to be', ' 'to have been', and ' (often shortened to ') 'to be going to be'. Other irregular present infinitives are ' (sometimes in Plautus and Lucretius ') 'to be able', and '/' 'to eat'. \n", "Now is often used at the end of sentences or phrases as a semantically empty word, completing an utterance without contributing any apparent meaning. Examples include \"Bye now\" (= \"Goodbye\"), \"There you go now\" (when giving someone something), \"Ah now!\" (expressing dismay), \"Hold on now\" (= \"wait a minute\"), \"Now then\" as a mild attention-getter, etc. This usage is universal among English dialects, but occurs more frequently in Hiberno-English. It is also used in the manner of the Italian 'prego' or German 'bitte', for example a barman might say \"Now, Sir.\" when delivering drinks.\n", "BULLET::::- \"I'm coming\" can mean numerous things. Usually, it means \"I'm leaving and coming back within 5 minutes or not at all\". Whereas \"I'm coming now now\" means \"I'm coming right back now for sure\".\n", "The \"going to\" of this future construction is frequently contracted in colloquial English to \"gonna\", and in some forms of English the copula may also be omitted. Hence \"You're going to like it\" could be said as \"You're gonna like it\" or just \"You gonna like it\". In the first person, \"I'm gonna\" may further contract to \"I'm'n'a\" or \"I'mma\" , or frequently . (For derived forms found in English-based creole languages, see below.) This is true even when the main verb is elided, as in \"Yes, I'm/you're/etc. gonna (do that).\"\n", "BULLET::::- \"Like\" as a discourse marker. \"Like\" is used as a filler word, similar to \"um\" or \"er,\" as in, \"I'm, like, about to call my friend.\" It does not add content to the sentence, instead allowing time for the speaker to formulate what they will say next. The word is always unstressed when used in this way. This discourse marker is usually used to introduce quoted speech. For example, a person can start a conversation with stating, \"So, um, I'm like \"Where did he go? and she was um, like, 'I don't know, I haven't seen him.'\n", "Very often, different words can be used to alter the emphasis of a sentence – e.g. while and both mean \"I smoke, but I shouldn't\", the first sounds more like a statement of fact (\"...but I mustn't\"), while the second feels more like a \"judgement\" (\"...but I oughtn't\"). Similarly, and both mean \"I don't want to, but he does\", however the first emphasizes the fact that \"he\" wants to, while the second emphasizes the \"wanting\" rather than the person.\n", "Contractions formed from a pronoun and a form of \"to be\" are also not used, including: I'm, you're, we're, they're, he's, she's, it's, there's, here's, where's, how's, what's, who's, and that's. E-Prime also prohibits contractions of \"to be\" found in nonstandard dialects of English, such as \"ain't\".\n" ]
Do arteries unclog eventually? Or is plaque build up permanent?
HDL participates in a process called [reverse cholesterol transport](_URL_0_) that can remove ~~plaque~~ cholesterol from peripheral tissue. As qxrt points out, this might not remove the plaque itself.
[ "The progressive accumulation of plaque within the artery wall over decades is the setup for vulnerable plaque which, in turn, leads to heart attack and stenosis (narrowing) of the artery (known as coronary artery lesions). IVUS is of use to determine both plaque volume within the wall of the artery and/or the degree of stenosis of the artery lumen. It can be especially useful in situations in which angiographic imaging is considered unreliable; such as for the lumen of ostial lesions or where angiographic images do not visualize lumen segments adequately, such as regions with multiple overlapping arterial segments. It is also used to assess the effects of treatments of stenosis such as with hydraulic angioplasty expansion of the artery, with or without stents, and the results of medical therapy over time.\n", "Blockage of arteries due to plaque buildup may also result in ischemia. Even a small amount of plaque build up can result in the narrowing of passageways, causing that area to become more prone to blood clots. Large blood clots can also cause ischemia by blocking blood flow.\n", "Atherosclerosis causes plaque to form within the carotid artery walls, usually at the fork where the common carotid artery divides into the internal and external carotid artery. The plaque build up can narrow or constrict the artery lumen, a condition called stenosis. Rupture of the plaque can cause the formation of a blood clot in the artery. A piece of the formed blood clot often breaks off and travels (embolizes) up through the internal carotid artery into the brain, where it blocks circulation, and can cause death of the brain tissue, a condition referred to as ischemic stroke.\n", "Prior to a myocardial infarction (MI or heart attack), plaque may accumulate in the coronary arteries, Some plaque formations may rupture, causing a mechanical dislodgment of endothelial cells creating CEC. The plaque that stays lodged in the coronary arteries may restrict blood flow to the cardiac muscle. This causes ischemia; the progressive death of cardiac muscle due to lack of oxygen. If the heart muscles have prolonged ischemia this may lead to the detachment of endocardial endothelial cells which can also create CEC.\n", "If the muscular wall enlargement eventually fails to keep up with the enlargement of the atheroma volume, or a clot forms and organizes over the plaque, then the lumen of the artery becomes narrowed as a result of repeated ruptures, clots & fibrosis over the tissues separating the atheroma from the blood stream. This narrowing becomes more common after decades of living, increasingly more common after people are in their 30s to 40s.\n", "The plaque can be stable and asymptomatic, or it can be a source of embolization. Emboli break off from the plaque and travel through the circulation to blood vessels in the brain. As the vessels get smaller, an embolus can lodge in the vessel wall and restrict the blood flow to parts of the brain. This ischemia can either be temporary, yielding a transient ischemic attack, or permanent resulting in a thromboembolic stroke.\n", "Typically, coronary artery disease occurs when part of the smooth, elastic lining inside a coronary artery (the arteries that supply blood to the heart muscle) develops atherosclerosis. With atherosclerosis, the artery's lining becomes hardened, stiffened, and accumulates deposits of calcium, fatty lipids, and abnormal inflammatory cells – to form a plaque. Calcium phosphate (hydroxyapatite) deposits in the muscular layer of the blood vessels appear to play a significant role in stiffening the arteries and inducing the early phase of coronary arteriosclerosis. This can be seen in a so-called metastatic mechanism of calciphylaxis as it occurs in chronic kidney disease and hemodialysis (Rainer Liedtke 2008). Although these people suffer from a kidney dysfunction, almost fifty percent of them die due to coronary artery disease. Plaques can be thought of as large \"pimples\" that protrude into the channel of an artery, causing a partial obstruction to blood flow. People with coronary artery disease might have just one or two plaques, or might have dozens distributed throughout their coronary arteries. A more severe form is \"chronic total occlusion\" (CTO) when a coronary artery is completely obstructed for more than 3 months.\n" ]
Can anyone please explain the difference between LUMENS and candela per square meter (CD/sqm) (In human language) as a measure of light brightness?
A lumen gives the *total* amount of human-visible light given out. How bright the light actually is will depend on how wide the beam is. A very wide beam with a lot of lumens is as bright as a very narrow beam with a few lumens. A candela is maybe a more useful unit here. A candela tells you how intense the human-visible light is in every direction. It's independent of the size of the beam - it's basically the brightness "per beam size", if that makes sense. A 1 candela beam with a very small beam width looks just as bright as a 1 candela beam with a very large beam width. The both illuminate with the same brightness, it's just that one covers a bigger area than the other. So it's maybe a better unit for a headlight. Though it might not matter if all lights have similar beam-widths. However, "candelas per square metre" is a bit odd. As a unit, it makes sense, and it's something that we use in astronomy. It means something like "the brightness in a specific direction per square metre of source". That is, it's useless unless you actually know the size of the light source. I don't think that's the intent - it's probably mislabelled, and it really means just candelas. Given that, it's quite likely that they're using the units incorrectly. So I wouldn't completely trust when they say "lumens" that they don't actually mean "lumens". Which isn't super helpful, I know.
[ "The lumen is defined as amount of light given into one steradian by a point source of one candela strength; while the candela, a base SI unit, is defined as the luminous intensity of a source of monochromatic radiation, of frequency 540 terahertz, and a radiant intensity of 1/683 watts per steradian. (540 THz corresponds to about 555 nanometres, the wavelength, in the green, to which the human eye is most sensitive. The number 1/683 was chosen to make the candela about equal to the standard candle, the unit which it superseded).\n", "The lumen can be thought of casually as a measure of the total \"amount\" of visible light in some defined beam or angle, or emitted from some source. The number of candelas or lumens from a source also depends on its spectrum, via the nominal response of the human eye as represented in the luminosity function.\n", "The lumen (lm) is the SI unit for luminous flux, the total amount of light emitted by a source, weighted according to the sensitivity of the human eye to various colours of light. Lumens per watt is a common measure of the efficacy of a light source. The luminous flux is of less value for bicycle lighting due to the importance of directionality. Luminous intensity is much more useful, but lumens per watt is a handy way to compare the output of otherwise similar lights.\n", "The lumen (symbol: lm) is the SI derived unit of luminous flux, a measure of the total quantity of visible light emitted by a source per unit of time. Luminous flux differs from power (\"radiant flux\") in that radiant flux includes all electromagnetic waves emitted, while luminous flux is weighted according to a model (a \"luminosity function\") of the human eye's sensitivity to various wavelengths. Lumens are related to lux in that one lux is one lumen per square meter.\n", "While luma is more often encountered, relative luminance is sometimes used in video engineering when referring to the brightness of a monitor. The formula used to calculate relative luminance uses coefficients based on the CIE color matching functions and the relevant standard chromaticities of red, green, and blue (e.g., the original NTSC primaries, SMPTE C, or Rec. 709). For the Rec. 709 (and sRGB) primaries, the linear combination, based on pure colorimetric considerations and the definition of relative luminance is:\n", "The peak of the luminosity function is at 555 nm (green); the eye's image-forming visual system is more sensitive to light of this wavelength than any other. For monochromatic light \"of this wavelength\", the amount of illuminance for a given amount of irradiance is maximum: 683.002 lux per 1 W/m; the irradiance needed to make 1 lux at this wavelength is about 1.464 mW/m. Other wavelengths of visible light produce fewer lux per watt-per-meter-squared. The luminosity function falls to zero for wavelengths outside the visible spectrum.\n", "The lumen method in brief consists of calculation of the \"cavity ratios\" of the upper, middle, and lower volumes of the space to be lighted. The lower cavity is from the floor to the working height, the upper cavity is from the lower edge of the luminaires to the ceiling, and the middle cavity is the volume between these planes.\n" ]
What were the predominant armor types worn in ancient southwest Asia?
The most common armour was scale body armour with a helmet of approximately conical or hemispherical shape. Surviving scales are commonly bronze, but bronze has an advantage in survival over rawhide and iron, which were also used. Descriptions of metal armour often describe it "gold" or "silver", which might refer to bronze and iron, respectively. Scale body armour could be either short, to the waist or hips, or long, covering to about knee level. Most armours appear to have been sleeveless or with short sleeves. Next most common would be lamellar, either short or long, and sleeveless or short-sleeved, like scale armours. Again, surviving lamellae are often bronze, but rawhide and iron were used too. Mail probably appears in the region in late antiquity, and was in use by the early 3rd century AD. The first major users of mail in the area appear to be the Sassanids. There are some armours with long sleeves for the arms, notably cataphract armours. Scale sleeves were used, and art suggest lamellar sleeves. Sassanid armoured cavalry would typically have long mail sleeves. Horse armour was also in use. I don't recall seeing mail horse armour, but both scale and lamellar horse armour were used. There was some use of plate armour, from the mid/late first millenium BC. Body armour included round plates used to protect the torso, and more complete plate cuirasses, but these seem to be much less common than scale/lamellar/mail which appear much more often in art. Plate greaves were used; this was quite likely Greek/Hellenistic influence. H. Russell Robinson, "Oriental Armour" covers antiquity in the first chapter.
[ "Protective clothing and armour have been used by armies from earliest recorded history; the King James Version of the Bible [Jeremiah 46:4] translates the Hebrew סריון \"ÇiRYON\" or שריון \"SiRYoN\" \"coat of mail\" as \"brigandine\". Medieval brigandines were essentially a refinement of the earlier coat of plates, which developed in the late 12th century, typically of simpler construction made of larger plates. The Asian-originated armour reached Europe after the Mongol invasion in 1240 that destroyed the Kievan Rus' and generated extensive damage to the Kingdom of Hungary in 1241. The new armour became very popular first in Eastern Europe, especially in Hungary, towards the end of the 13th century and after having proved effective was adopted by the medieval states from West Europe several decades later.\n", "Less known examples were present in Asia from Iran to Mongolia, including Central Asia. Laminar armor from animal skins has also been traditionally made and worn in the Arctic areas of what are now Siberia, Alaska and Canada.\n", "In East Asia many types of armour were commonly used at different times by various cultures, including scale armour, lamellar armour, laminar armour, plated mail, mail, plate armour and brigandine. Around the dynastic Tang, Song, and early Ming Period, cuirasses and plates (mingguangjia) were also used, with more elaborate versions for officers in war. The Chinese, during that time used partial plates for \"important\" body parts instead of covering their whole body since too much plate armour hinders their martial arts movement. The other body parts were covered in cloth, leather, lamellar, or Mountain pattern. In pre-Qin dynasty times, leather armour was made out of various animals, with more exotic ones such as the rhinoceros.\n", "Early types of this armour were known among the Celtiberians, by the Romans, in the Middle East, Central Asia, India, Russia, Siberia (where it was worn by Siberian natives before the Russian conquest), Mongolia, Indochina and China (including Tibet too).\n", "Eastern armor has a long history, beginning in Ancient China. In East Asian history laminated armor such as lamellar, and styles similar to the coat of plates, and brigandine were commonly used. Later cuirasses and plates were also used. In pre-Qin dynasty times, leather armor was made out of rhinoceros. Chinese influence in Japan would result in the Japanese adopting Chinese styles, their samurai armor being a result of this influence.\n", "According to Bobrov, until the end of the 15th century the most popular armour in certain regions including Central Asia and Iran was lamellar armour, brigandines, and laminar armour. However, in Iran since the 15th century lamellar and laminar armour were typical only in the south, while during the 15th century the typical armour in the north was plated mail.\n", "The best preserved armors from the three kingdoms period originate almost exclusively from the Gaya confederacy. The armour from Gaya are the best examples of plate armor from ancient times rivaling those of Mediterranean basin from the same period. These Gaya style plate armour are categorized into three types- one is made by joining vertical steel bands to form a single plate, another by joining horizontal bands, and the other by putting small triangular steel pieces together. The first type is found in Gaya and Silla, while most examples for the other two are found in Gaya but some have been found in northern Baekje. Similar styles have been also found in Kyushu and Honsu, Japan.\n" ]
why does raw seafood smell so much stronger than other animal proteins?
Chemicals in fish digestive systems that break down food (ie: other fish) turn on the fish's own body when it dies, speeding up decomposition. This is why you're supposed to gut a fish asap. The chemicals that cause the odor itself come from the decomposition and are called amines. _URL_0_
[ "Seafood, however, would logically have a more distinctive flavor. (The extent of its divergence is not consistent; tuna was said to taste enough like chicken that a prominent tuna canner named its product Chicken of the Sea.) Also, although mammals are tetrapods, very few mammals taste like chicken, which implies that there had been a mutation that changed their flavor on that branch of the evolutionary tree.\n", "During the production of surströmming, just enough salt is used to prevent the raw herring from rotting while allowing it to ferment. A fermentation process of \"at least six months\" gives the fish its characteristic strong smell and somewhat acidic taste. According to a Japanese study, a newly opened can of surströmming has one of the most putrid food smells in the world, even stronger than similarly fermented fish dishes such as the Korean hongeohoe or Japanese kusaya.\n", "Fish and seafood — such as mackerel, mackerel tuna, tuna, skipjack tuna, red snapper, gourami, pangasius, hemibagrus or cuttlefish — either the whole body or sometimes only the fish heads are added to make a spicy and tart fish stew. It is important that the fish remain intact for serving so generally the fish is added last.\n", "Unprocessed shark meat may have a strong odor of ammonia, due to the high urea content that develops as the fish decomposes. The urea content and ammonia odor can be reduced by marinating the meat in liquids such as lemon juice, vinegar, milk, or saltwater. Preparation methods include slicing the meat into steaks and fillets.\n", "Over 33,000 species of fish and many more marine invertebrate species have been described. Bromophenols, which are produced by marine algae, gives marine animals an odor and taste that is absent from freshwater fish and invertebrates. Also, a chemical substance called dimethylsulfoniopropionate (DMSP) that is found in red and green algae is transferred to animals in the marine food chain. When broken down, dimethyl sulfide (DMS) is produced, and is often released during food preparation when fresh fish and shellfish are heated. In small quantities it creates a specific smell one associates with the ocean, but which in larger quantities gives the impression of rotten seaweed and old fish. Another molecule known as TMAO occurs in fishes and give them a distinct smell. It also exists in freshwater species, but becomes more numerous in the cells of an animal the deeper it lives, so that fish from the deeper parts of the ocean has a stronger taste than species who lives in shallow water. Eggs from seaweed contains sex pheromones called dictyopterenes, which are meant to attract the sperm. These pheromones are also found in edible seaweeds, which contributes to their aroma. However, only a small number of species are commonly eaten by humans.\n", "Food technologists have long known that protein hydrolysis produces a meat bouillon-like odor and taste. In 1831, Berzelius obtained products having a meat bouillon taste when hydrolysing proteins with hydrochloric acid.\n", "In animals such as dogs and cats, protein maintains the health and quality of the skin by promoting hair follicle growth and keratinization, and thus reducing the likelihood of skin problems producing malodours. Poor-quality proteins also have a role regarding gastrointestinal health, increasing the potential for flatulence and odorous compounds in dogs because when proteins reach the colon in an undigested state, they are fermented producing hydrogen sulfide gas, indole, and skatole. Dogs and cats digest animal proteins better than those from plants but products of low-quality animal origin are poorly digested, including skin, feathers, and connective tissue.\n" ]
When we mix two colors, do we create a new pigment, or are the colors still separate, just impossible to distinguish?
Tiny spots of yellow and green. Paint mixtures aren't chemical reactions, there is no bonding. I mean, I can't speak to all paints, but oil, watercolour, and acrylic at least are all distinct pigments suspended in various solutions.
[ "In this traditional scheme, a complementary color pair contains one primary color (yellow, blue or red) and a secondary color (green, purple or orange). The complement of any primary color can be made by combining the two other primary colors. For example, to achieve the complement of yellow (a primary color) one could combine red and blue. The result would be purple, which appears directly across from yellow on the color wheel. Continuing with the color wheel model, one could then combine yellow and purple, which essentially means that all three primary colors would be present at once. Since paints work by absorbing light, having all three primaries together produces a black or gray color (see subtractive color). In more recent painting manuals, the more precise subtractive primary colors are magenta, cyan and yellow.\n", "One reason the artist's primary colors work at all is due to the imperfect pigments being used have sloped absorption curves, and change color with concentration. A pigment which is pure red at high concentrations can behave more like magenta at low concentrations. This allows it to make purples that would otherwise be impossible. Likewise, a blue that is ultramarine at high concentrations appears cyan at low concentrations, allowing it to be used to mix green. Chromium red pigments can appear orange, and then yellow, as the concentration is reduced. It is even possible to mix very low concentrations of the blue mentioned and the chromium red to get a greenish color. This works much better with oil colors than it does with watercolors and dyes.\n", "The additive mixing of colors is not commonly taught to children, as it does not correspond to the mixing of physical substances (such as paint) which would correspond to subtractive mixing. Two beams of light that are superimposed mix their colors additively.\n", "Real colors are colors that can be produced by a physical light source. Any additive mixture of two real colors is also a real color. When colors are displayed in the CIE 1931 XYZ color space, additive mixture results in a color along the line between the colors being mixed. By mixing any three colors, one can therefore create any color contained in the triangle they describe — this is called the gamut formed by those three colors, which are called primary colors. Any colors outside of this triangle cannot be obtained by mixing the chosen primaries.\n", "Charles Blanc's Grammaire des arts du dessin introduced Seurat to the theories of color and vision that would inspire chromo-luminarism. Blanc's work, drawing from the theories of Michel Eugène Chevreul and Eugène Delacroix, stated that optical mixing would produce more vibrant and pure colors than the traditional process of mixing pigments. Mixing pigments physically is a subtractive process with cyan, magenta, and yellow being the primary colors. On the other hand, if colored light is mixed together, an additive mixture results, a process in which the primary colors are red, green and blue. The optical mixture which characterized Divisionism—the process of mixing color by juxtaposing pigments—is different from either additive or subtractive mixture, although combining colors in optical mixture functions the same way as additive mixture, i.e. the primary colors are the same. In reality, Seurat's paintings did not actually achieve true optical mixing; for him, the theory was more useful for causing vibrations of color to the viewer, where contrasting colors placed near each other would intensify the relationship between the colors while preserving their singular separate identity.\n", "By convention, the three primary colors in additive mixing are red, green, and blue. In the absence of light of any color, the result is black. If all three primary colors of light are mixed in equal proportions, the result is neutral (gray or white). When the red and green lights mix, the result is yellow. When green and blue lights mix, the result is a cyan. When the blue and red lights mix, the result is magenta. \n", "Pigments, such as inks and paint, display color by absorbing some wavelengths of light and reflecting the remainder. When pigments are combined, they absorb the combination of their colors, and reflect less. Thus, combining pigments results in a darker color. This is called subtractive color-mixing, as mixing pigments subtracts wavelengths from the light that is reflected.\n" ]
Would you feel yourself spinning in absolutely empty space? If not, how is this possible?
Since spinning is a constant acceleration (constant change in velocity vector) the fluids in your ears should be displaced, and thus you should still be able to feel that you are spinning and get nauseous. If the spin is slow enough, I'd imagine that with zero reference points and a nearly non-existant fluid displacement, you wouldn't know you are spinning.
[ "The most common general symptom of having the spins is described by its name: the feeling that one has the uncontrollable sense of spinning, although one is not in motion, which is one of the main reasons an intoxicated person may vomit. The person has this feeling due to impairments in vision and equilibrioception. Diplopia (double vision) or polyplopia are common, as well as the symptoms of motion sickness and vertigo.\n", "Scientists in a rotating box can measure the speed and direction of their rotation by measuring these fictitious forces. For example, Léon Foucault was able to show the Coriolis force that results from Earth's rotation using the Foucault pendulum. If Earth were to rotate many times faster, these fictitious forces could be felt by humans, as they are when on a spinning carousel.\n", "One way to mask the symptoms of the spins is to avoid staring at moving objects, such as people who are dancing or ceiling fans. Instead, it helps to stare at a non-moving object and slowly blink a few times. However, it will make things worse to keep one's eyes closed for an extended period. In minor cases of the spins, simply sitting alone in a quiet place or taking a walk is all it takes to make them subside.\n", "indicate that, at such low rotation speeds, few people would experience motion sickness due to coriolis forces acting on the inner ear. People would, however, be able to detect spinward and antispinward directions by turning their heads, and any dropped items would appear to be deflected by a few centimetres. The central axis of the habitat would be a zero-gravity region, and it was envisaged that recreational facilities could be located there.\n", "[6:16:37] If I turn, the whole thing seems enormously flat. I mean, if I get the feeling that if a man stood on those sands, I think he would sink into them [laughs]. Maybe that's where that liquid feeling comes from.\n", "Rotating devices such as centrifuges used in astronaut training and amusement park rides such as the Rotor, and the Gravitron can cause motion sickness in many people. While the interior of the centrifuge does not appear to move, one will experience a sense of motion. In addition, centrifugal force can cause the vestibular system to give one the sense that downward is in the direction away from the center of the centrifuge rather than the true downward direction.\n", "When a spinning body passes through air at right angles to its axis of spin, it experiences a sideways force in the third dimension. This Magnus effect was first demonstrated on a spinning cylinder by Gustav Magnus in 1872. If the \n" ]
How long does light last for?
> What is light, Light is what you perceive when a photon of light hits a photo-receptor protein in one of the cells in your eye. The protein gets energy from the light that prompts a conformaitonal change (it wiggles), and this wiggle causes a change in the rate-of-fire of the neurons in your eye. This signals ("tells") the brain that you have seen light. Light is made of little tiny particles called photons. All photons travel at the "speed of light", c, around 300,000,000 m/s. They travel at the speed of light from every frame of reference. So if you are moving towards the light, you still measure it moving at c. If you are moving away from the light, it is still moving at c, from your perspective. There are an infinite number of different photons that can be created, because each photon has a specific "energy". There are high-energy photons and low-energy photons, and everything in between. High-energy photons are dangerous, and can cause cancer (gamma rays, X rays). Low energy photons are pretty much harmless (radio waves, visible light). All the different energies of light lie on a "spectrum". Humans can only detect a narrow range of light energies, from red to purple. Red photons have less energy, green have an in-between amount, and purple has the most. Above purple is ultraviolet (higher energy). Below red is infrared (low energy). {Humans glow. We emit photons. But these photons are low-energy infrared photons, which humans can't detect. So we can't see ourselves glow. But snakes can detect infrared photons. So snakes can see us glow!} > what is dark? Dark is a human/animal perception. It occurs when no light hits the sensory apparatus (the eye). > Does it degrade in any way Yes and no. When it is travelling through empty space, light does not have an expiration date. We can look in the sky and see photons from 1 billion years ago. It is only when light hits and interacts with an object that it changes. Light can be absorbed and converted to heat. Or converted to energy (as in solar cells or plants doing photosynthesis).
[ "Multiples of the light-second can be defined, although apart from the light-year, they are more used in popular science publications than in research works. For example, a light-minute is 60 light-seconds, and the average distance between Earth and the Sun is 8.317 light-minutes.\n", "Lighting begins at the International Date Line, in the first time zone, remaining lit a period of one hour, with the next time zone lighting respectively, moving through each time zone as the Wave of Light circumnavigates the globe. The result is a continuous chain of light encompassing and spanning across the world and around the globe for a 24-hour period, illuminating the night in love and light in honor and remembrance of our children.\n", "The light-second is a unit of length useful in astronomy, telecommunications and relativistic physics. It is defined as the distance that light travels in free space in one second, and is equal to exactly .\n", "The light is above sea level and has a nominal range of and geographical range of . Its current light characteristic is a single flash of 0.07 seconds every 7.5 seconds with an eclipse of 7.43 seconds.\n", "The light-year is a unit of length used to express astronomical distances and measures about 9.46 trillion kilometres (9.46 x 10 km) or 5.88 trillion miles (5.88 x 10 mi). As defined by the International Astronomical Union (IAU), a light-year is the distance that light travels in vacuum in one Julian year (365.25 days). Because it includes the word \"year\", the term light-year is sometimes misinterpreted as a unit of time.\n", "Each of the 7 lights refers to a time between 30 seconds and 2 minutes, 30 seconds. The times are 30 seconds/45 seconds/1 minute/1 minute, 30 seconds/1 minute, 45 seconds/2 minutes/2 minutes, 30 seconds. This time refers to the length of time of each round.\n", "Each Tree, in turn, would give off light for seven hours (waxing to full brightness and then slowly waning again), with the ends of their cycles overlapping so that at one hour each of \"dawn\" and \"dusk\" soft gold and silver light would be given off together. In total, therefore, one \"day\" of first silver then gold light lasted twelve hours. \n" ]
Is it possible to have multiple earth like planets in a planetary system?
It really depends on the planets. Venus and Mars are both within the "Goldilocks zone" for Sol. Depending on who's math they are right on the edge though. Venus ended up with a runaway greenhouse effect and now has an acid atmosphere, Mars lost its magnetic field and was stripped of its atmosphere. At one point, both of those planets were quite similar to earth so it can be done. The safe zone isn't a single orbit, it is a rather wide area so you could easily sneak a couple planetary orbits in there. There are billions of stars out there, anything that can exist almost certainly does exist, we just haven't found it yet.
[ "Planets that orbit just one star in a binary pair are said to have \"S-type\" orbits, whereas those that orbit around both stars have \"P-type\" or \"circumbinary\" orbits. It is estimated that 50–60% of binary stars are capable of supporting habitable terrestrial planets within stable orbital ranges.\n", "Star systems have at most five colonizable planets, and a few have none. Players can colonize all solid planet types, while gas giants and asteroids can be made habitable with the planet construction technology. Colonizable planets vary in several ways, making some more desirable than others:\n", "Terrestrial planets in multiple star systems, those containing three or more stars, are not likely to have stable orbits in the long term. Stable orbits in binary systems take one of two forms: S-Type (satellite or circumstellar) orbits around one of the stars, and P-Type (planetary or circumbinary) orbits around the entire binary pair. Eccentric Jupiters may also disrupt the orbits of planets in habitable zones.\n", "Some planets orbit one member of a binary star system, and several circumbinary planets have been discovered which orbit around both members of binary star. A few planets in triple star systems are known and one in the quadruple system Kepler-64.\n", "In the past scientists thought that planets would be unlikely to form around binary stars. However, recent simulations indicate that planets are just as likely to form around binary star systems as single-star systems. Of the 3457 exoplanets currently known, 146 actually orbit binary star systems (and 39 orbit multiple star systems with 3 or more stars). Specifically, they orbit what are known as \"wide\" binary star systems where the two stars are fairly far apart (several AU). Tatooine appears to be of the other type — a \"close\" binary, where the stars are very close, and the planets orbit their common center of mass.\n", "In non circumbinary planets, if a planet's distance to its primary exceeds about one fifth of the closest approach of the other star, orbital stability is not guaranteed. Whether planets might form in binaries at all had long been unclear, given that gravitational forces might interfere with planet formation. Theoretical work by Alan Boss at the Carnegie Institution has shown that gas giants can form around stars in binary systems much as they do around solitary stars.\n", "While a number of binary star systems have been found to harbor extrasolar planets, such systems are comparatively rare compared to single star systems. Observations by the Kepler space telescope have shown that most single stars of the same type as the Sun have plenty of planets, but only one-third of binary stars do. According to theoretical simulations, even widely separated binary stars often disrupt the discs of rocky grains from which protoplanets form. On the other hand, other simulations suggest that the presence of a binary companion can actually improve the rate of planet formation within stable orbital zones by \"stirring up\" the protoplanetary disk, increasing the accretion rate of the protoplanets within.\n" ]
how come seedless cherries aren’t a thing like seedless grapes and oranges?
The "seed" in a cherry is called a pit, and if I'm not mistaken, it's where the juicy part of the cherry came from. Similar to dates and plums. In contrast, the seeds of grapes and oranges are intended to produce more grapevines/orange trees
[ "The fruit is a deep burgundy colour, with a sour flavour and is popular in jams. It is cultivated in small plantations. Due to infertile seeds it can only be propagated from cuttings or division. Hence all cultivated material is derived from clones of wild plants. Plants take at least six years to produce fruit. Some selections are heavy bearing.\n", "Heirloom seeds can be easily collected and usually almost all seeds will continue to show the traits of the original seed because this family of tomatoes almost always self-pollinates. Dramatic cross-pollination may occur in the presence of pollinating insects during flowering. Collecting heirloom seed is as easy as picking ripe tomatoes, chopping or mashing into a jar till less than half-full, filling with water, shaking from time to time and allowing to decompose for 1–6 days until seeds sink to the bottom, then rinsing until the seeds are clean, and drying. This decomposition is beneficial because it discourages transmission of diseases to the seed, the drying promotes better germination, and because the seeds are easier to separate when they are clean.\n", "While tomatoes have the genes for producing anthocyanins, these genes are typically not expressed in the fruit in most commercial varieties. The result is that the pigments are limited to leaves and stems, where they are not eaten and would not provide any health benefit. With this in mind, researchers at Oregon State University have produced blue tomatoes using conventional breeding techniques, cross-pollinating domestic tomatoes with wild varieties that possess the \"Anthocyanin fruit\" (Aft) allele. As of the 2012 growing season, Oregon State University developed blue tomato seeds became commercially available under the cultivar name \"Indigo Rose\". The blue color is produced mostly by the anthocyanin petunidin on the outside of the tomato where the fruit is exposed to direct sunlight. The shaded side of the fruit is green when unripe, red when ripe, and the inside is red or deep pink. The tomatoes are small, about 2 inches across, round, and grow in clusters of 6 to 8. Flavor is described as slightly acidic. The vines are said to be indeterminate but compact, and disease resistant. While the concentration of anthocyanins is still very low compared to other fruits like blueberries, the pigments do improve the resistance of the fruits to the fungus Botrytis cinerea.\n", "Unlike most sweet cherry varieties, sour cherries are self fertile or self pollenizing (sometimes inaccurately referred to as self-pollinating). Two implications of this are that seeds generally run true to the cultivar, and that much smaller pollinator populations are needed because pollen only has to be moved within individual flowers. In areas where pollinators are scarce, growers find that stocking beehives in orchards improves yields.\n", "All cultivars of raspberries have perennial roots but, many do not have perennial shoots. In fact, most raspberries have shoots that are biennial (meaning shoots grow in the first growing season and fruits grow off of those shoots during the second growing season). The flowers can be a major nectar source for honeybees and other pollinators.\n", "Each grape berry contains a pedicel which attaches to the rachis. The main function of the rachis is to allow the grapes receive their water and nutrients. The pollination and fertilization of grapes results in one to four seeds within the berry. When fertilization does not occur, seedless grapes are formed, which are sought after for the production of raisins. Regardless of pollination and fertilization, most plants will produce around 100 to 200 grapes.\n", "Strawberries may also be propagated by seed, though this is primarily a hobby activity, and is not widely practiced commercially. A few seed-propagated cultivars have been developed for home use, and research into growing from seed commercially is ongoing. Seeds (achenes) are acquired either via commercial seed suppliers, or by collecting and saving them from the fruit.\n" ]
The maya treated the lower classes extremely poorly for over 2000 years, why was there no successful rebellion?
Can you elaborate? In what way do you think the Maya lower class was mistreated (and consistently for so long at that)? Is there some particular piece of evidence that led you to that conclusion?
[ "By the 9th and 10th centuries, this resulted in collapse of the system of rulership based around the divine power of the ruling lord. In the northern Yucatán, individual rule was replaced by a ruling council formed from elite lineages. In the southern Yucatán and central Petén, kingdoms generally declined; in western Petén and some other areas, the changes were catastrophic and resulted in the rapid depopulation of cities. Within a couple of generations, large swathes of the central Maya area were all but abandoned. Relatively rapid collapse affected portions of the southern Maya area that included the southern Yucatán Peninsula, northern Chiapas and Guatemala, and the area around Copán in Honduras. The largest cities had populations numbering 50,000 to 120,000 and were linked to networks of subsidiary sites. Both the capitals and their secondary centres were generally abandoned within a period of 50 to 100 years.\n", "It has been hypothesized that the decline of the Maya is related to the collapse of their intricate trade systems, especially those connected to the central Mexican city of Teotihuacán. Preceding improved knowledge of the chronology of Mesoamerica, Teotihuacan was believed to have fallen during 700–750, forcing the \"restructuring of economic relations throughout highland Mesoamerica and the Gulf Coast\". This remaking of relationships between civilizations would have then given the collapse of the Classic Maya a slightly later date. However, after knowing more about the events and the periods when they occurred, it is believed that the strongest Teotihuacan influence was during the 4th and 5th centuries. In addition, the civilization of Teotihuacan started to lose its power, and maybe abandoned the city, during 600–650. This differs greatly from the previous belief that Teotihuacano power decreased during 700–750. But since the new decline date of 600–650 has been accepted, the Maya civilizations are now thought to have lived on and prospered \"for another century and more\" than what was previously believed. Rather than the decline of Teotihuacan directly preceding the collapse of the Maya, their decline is now seen as contributing \"to the 6th-century 'hiatus'\".\n", "It may be argued in favor of a greater unity in the Southern Maya Area than the ethnic and linguistic diversity might otherwise indicate simply by virtue of the fact that a “Preclassic collapse” occurred extending through much of the Southern Maya Area . In the Southern Maya Area, in times called Classic for the Maya in the Lowlands to the north, tantalizing evidence exists of an abhorrence of a vacuum in the materially very rich breadbasket of the Area – and, as mentioned, particularly of a continuation of what must have been an extraordinary intensively cultivated commodity of enormous importance in Mesoamerica, and the Maya, in cuisine, ideologically, and even as currency, cacao. In the Guatemalan piedmont, located not more than sixty kilometers east of Chocola, Cotzumalguapa, of Middle Classic trajectory, is renowned for carved stone sculpture intimately associating decapitation and other sacrifice with cacao, associations we must conclude are representative of fierce warfare over this commodity, and copious ethnohistory from early after the Spanish Conquest makes reference to “chiefs” and chiefdoms fighting over production and distribution of the chocolate bean and/or its processed forms.\n", "What happened to Classic Maya civilization? Why did the entire society collapse—region wide—during a short crisis in the mid-ninth century A.D.? Was it persistent warfare? Agricultural or environmental collapse? Overpopulation? Drought? Or some combination . . . of some other factors altogether? Because of the extent of the Blue Creek site (from Middle Preclassic to the Terminal Classic collapse) and because of the length of time and extent of our studies, MRP is in a unique position to analyze a wealth of data that may provide critical insights into this mystery. This is an overarching focus of research over the coming years, into which all of our new projects will feed. This is at the vanguard of scholarship, impacting the debate about the Classic Maya collapse among scholars and institutions around the world.\n", "The Caste War of Yucatán, the Maya uprising that began in 1847, was one of the most successful modern Native American revolts. Maya rebels, or Cruzob, maintained relatively independent enclaves in the peninsula until the 1930s.\n", "Maya society underwent a series of profound transformations between c. 100 AD and 250 AD, which resulted in the cessation of monumental building at many Preclassic cities and the inferred collapse of their political and economic systems, often characterized as the \"Preclassic Collapse.\"\n", "Anthropologist Joseph Tainter wrote extensively about the collapse of the Southern Lowland Maya in his 1988 study \"The Collapse of Complex Societies\". His theory about Maya collapse encompasses some of the above explanations, but focuses specifically on the development of and the declining marginal returns from the increasing social complexity of the competing Maya city-states. Psychologist Julian Jaynes suggested that the collapse was due to a failure in the social control systems of religion and political authority, due to increasing socioeconomic complexity that overwhelmed the power of traditional rituals and the king's authority to compel obedience.\n" ]
Is temperature vary based on the point of reference?
Good question, and yes. We understand the asymptotic Hawking temperature, it becomes a constant for sufficiently distant observers at rest. In simplified units it takes on the form, T = k/2pi k = surface gravity at event horizon Let's look at Schwarzschild coordinates, ds^2 = -(1-R/r)(cdt)^2 + (dr)^2/(1-R/r) + angle terms R = Schwarzschild radius This corresponds to a frame which "hovers" above the black hole at a fixed radius, as you can see the temperature is modified by the curvature, T = a/2pi = (k/2pi)/√(1-R/r) * Unruh. ["Notes on black hole evaporation"](_URL_1_) Phys.Rev. D14 (1976) Here we see, the hovering temperature relies on the inverse of the square root of the curvature. There are two important features: 1. As you recede to flat space r-- > big, you recover the famous Hawking temperature. 2. As you approach the Event Horizon of the black hole, the temperature blows up to infinity. We can attribute this to the acceleration an observer must maintain to not fall into the black hole. What about free falling into the black hole? Like if I decided to jump in? Unfortunately, I could not find an exact mathematical treatment--except to say that there is an exact solution for 6-dimensional spacetime. * Brynjolfsson, Thorlacius. "Taking the Temperature of a Black Hole" JHEP 2008. _URL_3_ * Kim, Choi, Park. "Local free-fall Temperature of GMGHS Black Holes" Phys. Rev. D 89. 2013. _URL_2_ * _URL_0_ All these suggest that the free falling observer will see a finite radiation bath (which isn't precisely thermal) on the order of the Hawking temperature even as they cross the horizon--so they won't burn up!
[ "It makes good sense, for example, to say of the extensive variable , or of the extensive variable , that it has a density per unit volume, or a quantity per unit mass of the system, but it makes no sense to speak of density of temperature per unit volume or quantity of temperature per unit mass of the system. On the other hand, it makes no sense to speak of the internal energy at a point, while when local thermodynamic equilibrium prevails, it makes good sense to speak of the temperature at a point. Consequently, temperature can vary from point to point in a medium that is not in global thermodynamic equilibrium, but in which there is local thermodynamic equilibrium.\n", "For axiomatic treatment of thermodynamic equilibrium, since the 1930s, it has become customary to refer to a zeroth law of thermodynamics. The customarily stated minimalist version of such a law postulates only that all bodies, which when thermally connected would be in thermal equilibrium, should be said to have the same temperature by definition, but by itself does not establish temperature as a quantity expressed as a real number on a scale. A more physically informative version of such a law views empirical temperature as a chart on a hotness manifold. While the zeroth law permits the definitions of many different empirical scales of temperature, the second law of thermodynamics selects the definition of a single preferred, absolute temperature, unique up to an arbitrary scale factor, whence called the thermodynamic temperature. If internal energy is considered as a function of the volume and entropy of a homogeneous system in thermodynamic equilibrium, thermodynamic absolute temperature appears as the partial derivative of internal energy with respect the entropy at constant volume. Its natural, intrinsic origin or null point is absolute zero at which the entropy of any system is at a minimum. Although this is the lowest absolute temperature described by the model, the third law of thermodynamics postulates that absolute zero cannot be attained by any physical system.\n", "In some cases, the thermodynamic parameter is actually defined in terms of an idealized measuring instrument. For example, the zeroth law of thermodynamics states that if two bodies are in thermal equilibrium with a third body, they are also in thermal equilibrium with each other. This principle, as noted by James Maxwell in 1872, asserts that it is possible to measure temperature. An idealized thermometer is a sample of an ideal gas at constant pressure. From the ideal gas law, the volume of such a sample can be used as an indicator of temperature; in this manner it defines temperature. Although pressure is defined mechanically, a pressure-measuring device called a barometer may also be constructed from a sample of an ideal gas held at a constant temperature. A calorimeter is a device which is used to measure and define the internal energy of a system.\n", "Experimental physicists, for example Galileo and Newton, found that there are indefinitely many empirical temperature scales. Nevertheless, the zeroth law of thermodynamics says that they all measure the same quality. This means that for a body in its own state of internal thermodynamic equilibrium, every correctly calibrated thermometer, of whatever kind, that measures the temperature of the body, records one and the same temperature. For a body that is not in its own state of internal thermodynamic equilibrium, different thermometers can record different temperatures, depending respectively on the mechanisms of operation of the thermometers.\n", "For some other quantities, it is easier or has been convention to estimate ratios between attribute \"differences\". Consider temperature, for example. In the familiar everyday instances, temperature is measured using instruments calibrated in either the Fahrenheit or Celsius scales. What are really being measured with such instruments are the magnitudes of temperature differences. For example, Anders Celsius defined the unit of the Celsius scale to be 1/100th of the difference in temperature between the freezing and boiling points of water at sea level. A midday temperature measurement of 20 degrees Celsius is simply the ratio of the Celsius unit to the midday temperature.\n", "Temperature is a statistical quantity. The formal definition is T = dU/dS, the change in internal energy with respect to entropy, holding volume and particle number constant. A practical definition comes from the fact that the atoms, molecules, or whatever particles in a system have an average kinetic energy. The average means to average over the kinetic energy of all the particles in a system.\n", "As noted in the section above, accurate formulation for temperatures may require analysis based on changing heat transfer coefficients at different temperatures, a situation frequently found in free-convection situations, and which precludes accurate use of Newton's law.\n" ]
Why can't we run 64 bit applications in a 32 bit operating system, even if the CPU supports 64 bit?
A 32 bit operating system running on an x64 processor causes the processor to run in a 32-bit mode. For example, it uses 32-bit addressing in its virtual memory system. See for example: _URL_0_
[ "The 64-bit version runs on CPUs compatible with x86 8th generation (known as x86-64, or x64) or newer, and can run 32-bit and 64-bit programs. 32-bit programs and operating system are restricted to supporting only of memory while 64-bit systems can theoretically support of memory. 64-bit operating systems require a different set of device drivers than those of 32-bit operating systems.\n", "64-bit versions of Windows cannot run 16-bit software. However, most 32-bit applications will work well. 64-bit users are forced to install a virtual machine of a 16- or 32-bit operating system to run 16-bit applications.\n", "x86-based 64-bit systems sometimes lack equivalents of software that is written for 32-bit architectures. The most severe problem in Microsoft Windows is incompatible device drivers for obsolete hardware. Most 32-bit application software can run on a 64-bit operating system in a compatibility mode, also termed an emulation mode, e.g., Microsoft WoW64 Technology for IA-64 and AMD64. The 64-bit Windows Native Mode driver environment runs atop 64-bit NTDLL.DLL, which cannot call 32-bit Win32 subsystem code (often devices whose actual hardware function is emulated in user mode software, like Winprinters). Because 64-bit drivers for most devices were unavailable until early 2007 (Vista x64), using a 64-bit version of Windows was considered a challenge. However, the trend has since moved toward 64-bit computing, more so as memory prices dropped and the use of more than 4 GB of RAM increased. Most manufacturers started to provide both 32-bit and 64-bit drivers for new devices, so unavailability of 64-bit drivers ceased to be a problem. 64-bit drivers were not provided for many older devices, which could consequently not be used in 64-bit systems.\n", "Although 32-bit applications can be run transparently, the mixing of the two types of code within the same process is not allowed. A 64-bit program cannot use a 32-bit dynamic-link library (DLL) and similarly a 32-bit program cannot use a 64-bit DLL. This may lead to the need for library developers to provide both 32-bit and 64-bit binary versions of their libraries. Specifically, 32-bit shell extensions for Windows Explorer fail to work with 64-bit Windows Explorer. Windows XP x64 Edition ships with both 32-bit and 64-bit versions of Windows Explorer. The 32-bit version can become the default Windows Shell. Windows XP x64 Edition also includes both 32-bit and 64-bit versions of Internet Explorer 6, so that user can still use browser extensions or ActiveX controls that are not available in 64-bit versions.\n", "32-bit applications that include only 32-bit kernel-mode device drivers, or that plug into the process space of components that are implemented purely as 64-bit processes (e.g. Windows Explorer) cannot be executed on a 64-bit platform.\n", "Although 32-bit applications can be run transparently, the mixing of the two types of code within the same process is not allowed. A 64-bit application cannot link against a 32-bit library (DLL) and, similarly, a 32-bit application cannot link against a 64-bit library. This may lead to the need for library developers to provide both 32- and 64-bit binary versions of their libraries. Windows XP x64 Edition includes both 32- and 64-bit versions of Internet Explorer 6, in order to allow for the possibility that some third-party browser plugins or ActiveX controls may not yet be available in 64-bit versions.\n", "On 64-bit hardware with x86-64 architecture (AMD64), most 32-bit operating systems and applications can run with no compatibility issues. While the larger address space of 64-bit architectures makes working with large data sets in applications such as digital video, scientific computing, and large databases easier, there has been considerable debate on whether they or their 32-bit compatibility modes will be faster than comparably priced 32-bit systems for other tasks.\n" ]
Why do rivers meander?
I'm not an expert but... Water takes the path of least resistance. So if the ground is lower in one spot vs another, that is the direction the river will take. Or if the ground is softer or more prone to erosion, that will help to guide the path of the river. Once the river is bent, the erosion will be strongest on the outside of the curve, which will cause the curves to become more and more pronounced over time.
[ "Meanders are sinuous bends in rivers or other channels, which form as a fluid, most often water, flows around bends. As soon as the path is slightly curved, the size and curvature of each loop increases as helical flow drags material like sand and gravel across the river to the inside of the bend. The outside of the loop is left clean and unprotected, so erosion accelerates, further increasing the meandering in a powerful positive feedback loop.\n", "Meandering rivers flow higher and hence with more total flow, pressure and erosion on the outside of their bends due to forming a vortex as in a stirred coffee cup and consequently the river erodes more the outer bank. On the inside bend of a river, the level is lower, secondary flow moves sand and gravel across the river bed creating shallows and point bars, and friction of air and perturbances of the bed act against a higher proportion of the column of water, being shorter, slowing the water to varying degrees. Rivers are commonly described and interpreted by their sinuosity. The term is equally used to describe the actual incidence of and potential tendency of a river to curve or meander over its length. It is expressed as the ratio of the distance between two distant points in a river following the middle-of-the-river course of the river as compared with the straight distance between those points. Three conventional categorizations of rivers or their reaches exist. Meandering rivers a sinuosity value/ratio of greater than 1.5. A sinuosity value of less than 1.1 is a “straight” river. Between these values, a river is described as sinuous which describes those in a transitory state between the two states. Braided rivers do not follow this same convention. Meandering rivers trend in the direction of increasing sinuosity.\n", "The meanders that form part of either an entrenched river or meander valley are most commonly known as \"incised meanders\". They are commonly classified as either an \"ingrown meander\" or an \"entrenched meander\". For a long time, it was argued that ingrown meander occurs when downcutting process is slow and the river can cause lateral erosion, leading to an asymmetric valley. In addition, it was also argued for a long time that an entrenched meander forms when there is a rapid incision of the river bed such that the river does not have the opportunity to erode the lateral side. This leads to symmetrical valleys with a gorge-like appearance. However, more detailed studies have shown the development of ingrown meanders versus entrenched meanders depends on a complex mixture of factors such as bedrock lithology, tectonic activity, and climate.\n", "Rivers flowing downhill, from river source to river mouth, do not necessarily take the shortest path. For alluvial streams, straight and braided rivers have very low sinuosity and flow directly down hill, while meandering rivers flow from side to side across a valley. Bedrock rivers typically flow in either a fractal pattern, or a pattern that is determined by weaknesses in the bedrock, such as faults, fractures, or more erodible layers.\n", "In a freely meandering river, a \"slip-off slope\" is characterized by a gentle slope composed of sand and pebbles on the inside convex bank of a meander loop, across the channel from a cut bank or river-cut cliff. As water in a meandering river travels around a bend, it moves in a secondary corkscrew-like flow as it travels downstream, in a pattern called helicoidal flow. This phenomenon causes increased water velocity in the outside bend of the meander, driving lateral bank erosion. It is also responsible for slower water velocity on the inner bend of the meander. This low velocity allows eroded sediments from the cut bank side to be deposited on the inner bank. The deposition of material at the toe of a slip-off slope often results in the formation of a point bar. Geologically, the environment in which a slip-off slope is deposited is considered to be one of low energy.\n", "A river constantly evolves and as it does, meanders that were once a part of the river are abandoned in favor of a route that is more efficient for a river to take. As these old meanders are cutoff from the rest of the river, a new channel, or cutoff channel, is formed. \n", "Meanders are looping changes of direction of a stream caused by the erosion and deposition of bank materials. These are typically serpentine in form. Typically, over time the meanders gradually migrate downstream.\n" ]
why is a vehicle released in (for example) 2018 called a 2019?
There isn’t really a singular reason rather a few different things that led to it. Ford started releasing the next year’s model in the fall for those in agriculture. That was when they had the most money so they released the model at a time that people could buy next year’s model instead of buying the almost year old current year model. Other companies followed suit and started releasing in the fall, after WWII they settled on Oct 1 for an industry standard. Jumping to the 60s this also coincided with the rising television industry’s start of the fall season which always sees a spike in viewership which meant more people seeing ads
[ "On 5 November 2013, Toyota USA announced the 2014 model year Trail Teams edition would be called the \"Ultimate Edition\" and that the 2014 model year would be the last for the FJ Cruiser in that market. It continued to be made for sale in other markets such as Australia until its export to that market was discontinued in August 2016. It is still sold in the Middle East as of August 2019.\n", "In February 2019, Toyota introduced a Heritage edition of the Land Cruiser at the Chicago Auto Show as part of the 60th anniversary of the introduction of the FJ40. Toyota will only build 1200 vehicles and they will only be available as a 5-seater sold in the United States. Sales are expected to begin in the third quarter of 2019 for the 2020 model year.\n", "Originally, 168 additional cars were proposed to be built and provided for service on the , , , and services between 2015 and 2019; the contract number for these growth cars was unknown, but they were not delivered prior to 2019. However, it is expected that these \"new\" cars could be cars displaced from other services after the R179s are delivered, as some of the R32s (which were originally planned to be replaced completely by the R179s) will not be replaced until the R211 order is delivered.\n", "vehicles. In September the issue looks at the latest cars (sedans, coupes, etc.) for the next coming calendar year, whilst in October the magazine looks at off-roaders, MPVs and sport-utility vehicles (SUVs).\n", "Due to the passage of the Low Volume Vehicle Manufacturing Act, DMC Texas announced that they would be producing replica DeLoreans with an expected release date in 2017. DMC anticipates to build approximately 50 vehicles per year over the next 6 years with an estimated retail price of US$100,000. In October 2016, DMC announced that they are expecting to build 12 units in the first production year with as many as 50 in the second year of production. In February 2018, it was reported that production should start in January 2019.\n", "The is a series of four-wheel drive vehicles produced by the Japanese automobile manufacturer Toyota. It is Toyota's longest running series of models. , the sales of the Land Cruiser totalled more than 6.5 million units worldwide.\n", "2014: Toyota introduced the 4-door wagon GRJ76 and double-cabin GRJ79 pickup body in the Japan market as a limited-edition '30th Anniversary' Series 70. This model came with the 1GR-FE V6 petrol engine and 5 speed manual transmission. This model was on limited sale in Japan only until June 2015 when Japanese law required all new vehicles to come with Vehicle Skid Control, something the 70 Series does not have.\n" ]
who controls the internet?
Short answer is "no" on both accounts. Longer answer is that yes, there is monitoring on many levels of the internet that is maintained by agencies and law enforcement as well as by NGOs and lobby organizations. A government agency is unlikely to care that you're torrenting, but will care if you're making terrorist plots. Law enforcement will likely not care about your torrents either, but will care if you run a pedophile ring, and NGOs such as the MPAA will very much care about your torrenting, but they ahve very limited access to the internet. The US government does not control the internet. It isn't unlikely that they exert greater control than they let on through surveillance , as indicated by the last few years' leaks, and them asking corporations to install backdoors in software. The US did not create the internet either, really. They _did_ create ARPAnet, which is a predecessor and have had a great hand in shaping and developing the internet, or at least universities have, but the greater question is "what is the internet", which makes things more complicated. The internet is many things. The internet is infrastructure - It is radio towers to distribute wifi and cables to run ethernet through, some of them transcontinental and transoceanic. The internet is hardware - It is routers and switches and modems and servers The internet is also software, with applications, sites, server software, DNS, storage, databases. Many of these layers and segments are distributed, and none of them are government controlled - Most are controlled by working groups and organizations that work with maintaining and developing standards, and more hands-on with private companies who manufacture and support the hardware and software which runs these things. To reconnect to your first question: Yes, there is monitoring by agencies and law enforcement to hinder illegal activities such as trade with drugs, weapons and people, wire fraud, financing or plotting terrorism etc, but it's not one single entity that monitors the entire internet for it, rather government agencies and law enforcement that all work according to the laws and interests of teir nation (or the EU, or NATO, or whatever), with collaboration where possible and necessary. This also involves a lot of real-world work as well, monitoring people, suspects, their activities, working to establish links between online personae and real-world people etc.
[ "Internet service providers in many countries are legally required (e.g., via Communications Assistance for Law Enforcement Act (CALEA) in the U.S.) to allow law enforcement agencies to monitor some or all of the information transmitted by the ISP, or even store the browsing history of users to allow government access if needed (e.g. via the Investigatory Powers Act 2016 in the United Kingdom). Furthermore, in some countries ISPs are subject to monitoring by intelligence agencies. In the U.S., a controversial National Security Agency program known as PRISM provides for broad monitoring of Internet users traffic and has raised concerns about potential violation of the privacy protections in the Fourth Amendment to the United States Constitution. Modern ISPs integrate a wide array of surveillance and packet sniffing equipment into their networks, which then feeds the data to law-enforcement/intelligence networks (such as DCSNet in the United States, or SORM in Russia) allowing monitoring of Internet traffic in real time.\n", "The Internet's power is said to lie in its removal of a government’s control of information. Online on the Internet, any individuals can publish anything, which allows citizens to circumvent the government’s official information sources. This has threatened governing regimes and lead to many censoring or cutting Internet service in times of crisis.\n", "The government maintains control over the Internet by requiring all websites to register with the Information Affairs Authority (IAA). Internet Service Providers (ISPs) are indirectly controlled through the Telecommunications Regulation Authority (TRA) under the pretext of helping to protect intellectual property. A resolution issued by the Ministry of Culture and Information in January 2009 requires all ISPs to install a website blocking software chosen by the Ministry. The IAA and select Ministries within the government can decide to block certain websites, which is then enforced by the ISP.\n", "The government controls domestic Internet servers and sporadically monitors Internet usage, but by the end of 2012 it apparently did not have the ability to block access to Web sites. Authorities have developed infrastructure to route all Internet traffic through a single gateway, enabling them to monitor and restrict content. However, they apparently had not utilized this increased capability as of the end of 2012. The law generally protects privacy, including that of mail, telephone, and electronic correspondence, but the government reportedly continues to violate these legal protections when there is a perceived security threat. Security laws allow the government to monitor individuals' movements and private communications, including via cell phones and e-mail.\n", "The internet is unrestricted by the government, and is used by around 97% of the population. Government has put in place rules that force telecommunication companies to block popular file sharing sites like The Pirate Bay and KickassTorrents.\n", "Internet surveillance is the monitoring of Internet data traffic, web access, and other online activity. This may include monitoring any Internet traffic including encrypted web browser traffic on Secure Socket Layer (SSL) connections, personal web-based email, and personal banking sites. Research done by American Management Association that nearly 30 percent of all employers in the United States monitor employee e-mails.\n", "The government controls domestic Internet servers and sporadically monitors Internet usage, but by the end of 2012 it apparently did not have the ability to block access to Web sites. Authorities have developed infrastructure to route all Internet traffic through a single gateway, enabling them to monitor and restrict content. However, they apparently had not utilized this increased capability as of the end of 2012. The National Internet Committee under the Prime Minister’s Office administers the Internet system. The office requires Internet service providers to submit quarterly reports and link their gateways to facilitate monitoring, but the government’s enforcement capability appears limited.\n" ]
What elements/compounds give Jupiter its colors?
**The short answer**: Hydrocarbon hazes, Rayleigh scattering, and some unknown compound. **The long answer**: For the primary whites and browns that cover most of the planet, you need to realize that almost everything you see when you look at Jupiter is ammonia clouds, which on their own are bright white. Some latitudes are regions of upwelling (zones), and have high ammonia cloud-tops, while other latitudes are regions of downwelling (belts), and have low ammonia cloud-tops. In between these high and low heights sits a thick brown hydrocarbon haze, very chemically similar to smog. The cloud-tops in the zones are sticking up above most of the haze and thus appear fairly white. The cloud-tops in the belts, though, lie below the haze layer, and thus appear colored brown by the overlying haze. For the occasional bluish regions seen just to the north and south of the equator, these are some of the rare cloud clearings that occur in very strong downwelling regions. We're actually peering through the ammonia top cloud layer, and perhaps even down through the ammonium hydrosulfide middle cloud layer and the bottom water cloud layer. So, in those regions we're looking at just clear air, which has the exact same color as it does one Earth, blue. This is entirely due to Rayleigh scattering, the same reason that Earth's sky is blue. Then there's the reds, notably in Jupiter's Great Red Spot, although also occasionally seen in another big vortex here and there. As of right now, we don't actually know what makes the Great Red Spot red - this is generally known as the Jovian chromophore problem. Since this color is only seen in very large vortices, it's believed to be caused by some mixture of compounds already present on the planet getting pushed very high in the atmosphere by these vortices. In three dimensions, the Great Red Spot is essentially shaped like a wedding cake, so the cloud-tops at the center of the spot are at very high altitudes where there's a lot more ultraviolet light. You can end up producing all kinds of odd substances through UV photochemistry of trace substances in the atmosphere, and the working hypothesis at this point is that it's some kind of imine.
[ "Spectroscopically, the Jupiter trojans mostly are D-type asteroids, which predominate in the outer regions of the asteroid belt. A small number are classified as P or C-type asteroids. Their spectra are red (meaning that they reflect more light at longer wavelengths) or neutral and featureless. No firm evidence of water, organics or other chemical compounds has been obtained . 4709 Ennomos has an albedo slightly higher than the Jupiter-trojan average, which may indicate the presence of water ice. Some other Jupiter Trojans, such as 911 Agamemnon and 617 Patroclus, have shown very weak absorptions at 1.7 and 2.3 μm, which might indicate the presence of organics. The Jupiter trojans' spectra are similar to those of the irregular moons of Jupiter and, to a certain extent, comet nuclei, though Jupiter trojans are spectrally very different from the redder Kuiper belt objects. A Jupiter trojan's spectrum can be matched to a mixture of water ice, a large amount of carbon-rich material (charcoal), and possibly magnesium-rich silicates. The composition of the Jupiter trojan population appears to be markedly uniform, with little or no differentiation between the two swarms.\n", "The orange and brown coloration in the clouds of Jupiter are caused by upwelling compounds that change color when they are exposed to ultraviolet light from the Sun. The exact makeup remains uncertain, but the substances are thought to be phosphorus, sulfur or possibly hydrocarbons. These colorful compounds, known as chromophores, mix with the warmer, lower deck of clouds. The zones are formed when rising convection cells form crystallizing ammonia that masks out these lower clouds from view.\n", "There are many differences between SDOs and JFCs, even though many of the Jupiter-family comets may have originated in the scattered disc. Although the centaurs share a reddish or neutral coloration with many SDOs, their nuclei are bluer, indicating a fundamental chemical or physical difference. One hypothesis is that comet nuclei are resurfaced as they approach the Sun by subsurface materials which subsequently bury the older material.\n", "Sulfur, usually as sulfide, is present in many types of meteorites. Ordinary chondrites contain on average 2.1% sulfur, and carbonaceous chondrites may contain as much as 6.6%. It is normally present as troilite (FeS), but there are exceptions, with carbonaceous chondrites containing free sulfur, sulfates and other sulfur compounds. The distinctive colors of Jupiter's volcanic moon Io are attributed to various forms of molten, solid, and gaseous sulfur.\n", "Most polyhalogen ions are intensely colored, with deepened color as the atomic weight of the constituent element increases. The well-known starch-iodine complex has a deep blue color due to the linear ions present in the amylose helix. Some colors of the common species were listed below:\n", "Jupiter trojans are dark bodies with reddish, featureless spectra. No firm evidence of the presence of water, or any other specific compound on their surface has been obtained, but it is thought that they are coated in tholins, organic polymers formed by the Sun's radiation. The Jupiter trojans' densities (as measured by studying binaries or rotational lightcurves) vary from 0.8 to 2.5 g·cm. Jupiter trojans are thought to have been captured into their orbits during the early stages of the Solar System's formation or slightly later, during the migration of giant planets.\n", "The golden colour of caesium comes from the decreasing frequency of light required to excite electrons of the alkali metals as the group is descended. For lithium through rubidium this frequency is in the ultraviolet, but for caesium it enters the blue–violet end of the spectrum; in other words, the plasmonic frequency of the alkali metals becomes lower from lithium to caesium. Thus caesium transmits and partially absorbs violet light preferentially while other colours (having lower frequency) are reflected; hence it appears yellowish.\n" ]
how did spacex develop rocket technology so quickly?
It's not like they're starting with the same technology as nasa when it was founded.
[ "While SpaceX spent its own money to develop its first launch vehicle, the Falcon 1, the development of the Falcon 9 was accelerated by the purchase of several demonstration flights by NASA. This started with seed money from the Commercial Orbital Transportation Services (COTS) program in 2006. SpaceX was selected from more than twenty companies that submitted COTS proposals. Without the NASA money, development would have taken longer, Musk said.\n", "Since the founding of SpaceX in 2002, the company has developed three families of rocket engines — Merlin and the retired Kestrel for launch vehicle propulsion, and the Draco control thrusters. SpaceX is currently developing two further rocket engines: SuperDraco and Raptor. SpaceX is currently the world's most prolific producer of liquid fuel rocket engines.\n", "SpaceX is actively pursuing several different research and development programs. Most notable are those intended to develop reusable launch vehicles, an interplanetary transport system and a global telecommunications network.\n", "SpaceX utilizes a high degree of vertical integration in the production of its rockets and rocket engines. SpaceX builds its rocket engines, rocket stages, spacecraft, principal avionics and all software in-house in their Hawthorne facility, which is unusual for the aerospace industry. Nevertheless, SpaceX still has over 3,000 suppliers with some 1,100 of those delivering to SpaceX nearly weekly.\n", "Development work began in earnest before 2012 when SpaceX started to design the Raptor rocket engine which will propel all versions of the BFR launch vehicle (now renamed \"Super Heavy\") and spacecraft (now renamed \"Starship\"). Rocket engine development is one of the longest subprocesses in the design of new rockets.\n", "SpaceX has been explicit about its effort to bring iterative design practices into the space industry, and uses the technique on spacecraft, launch vehicles, electronics and avionics, and operational flight hardware operations.\n", "The SpaceX next-generation launch vehicle design combines several elements that, according to Musk, will make long-duration, beyond Earth orbit (BEO) spaceflights possible. The design is projected by SpaceX to reduce the per-ton cost of launches to low Earth orbit (LEO) and of transportation between BEO destinations. It will also serve all use cases for the conventional LEO market. This will allow SpaceX to focus the majority of their development resources on the next-generation launch vehicle.\n" ]
Why were the Turkish states able to assimilate Anatolia so quickly? Why didn't Persia or the Balkans assimilate?
I'll start off my answer by pointing out that significant parts of Iran and Afghanistan were in fact Turkified; Azerbaijan, parts of Fars, northeastern Iran, and northern and central Afghanistan. The main reason this happened, I would argue, is that the climate and habitat of these regions were (similarly to Anatolia) suitable to Turkish nomads. [I'm not too knowledgeable on Ottoman rule in the Balkans, so I'll leave that to someone else.[Though I would hazard a guess that Ottoman government was considerably more sedentary than their earlier Turkic colleagues and so were less concerned with the settlement of migrating Turkic tribes.]] The reason for this suitability lies in the mix of mountainous and flat terrain. The Turkic nomads who invaded Iran from Central Asia, and eventually moved further into Anatolia, practiced vertical transhumance. This involved the seasonal migration of the nomads' animal herds: down into the valleys/plains during the winter, and up into the lower reaches of the mountains during summer. The reason for the noticeably smaller degree of Turkification in Iran and Afghanistan compared to Anatolia is due to the presence in Iran/Afghanistan of competing groups. Kurds, Lurs, Pashtuns and many other Iranian groups already engaged in forms of nomadic pastoralism and so would have provided competition for space with the incoming Turks. Anatolia on the other hand, was populated mostly by urbanised, sedentary Greeks, Armenians, and Syriacs. There likely were some native nomadic groups, but certainly not to the same degree as Iran and Afghanistan. Consequently, the migrating Turks faced relatively little competition during their settlement of Anatolia. There were obviously other factors involved, such as the power vacuum in Anatolia after the Battle of Manzikert as the regional nobility fled to Constantinople to take part in the Byzantine intrigues (pun definitely intended) to select a new emperor, or in the remarkable ability of these Turkic tribes to assimilate Islamic and Christian ideas into their pre-existing pagan-influenced governing ideology. However, to reiterate it is the suitability of the terrain that I find to be the most important factor in the Turkification of these regions.
[ "In the resulting chaos, the Turks easily overran much of the Byzantine empire and, despite Byzantine reconquests and occasional western incursions in the form of crusading armies, a series of Turkish states were established in Anatolia. These Turkic tribes came around the south end of the Caspian Sea for the most part, and hence absorbed and transmitted Islamic culture and civilization in contrast to other Turks who, such as the Cumans, became partially Westernized and Christianized. With some superiority in population and organization, regional power naturally came to rest in the hands of the Turkic speaking population. Many Turkic people came into the region fleeing from the Mongol invasions, and others later came in as soldiers fighting in the ranks of Mongol armies. Turkic Islamized populations also absorbed large numbers of the older inhabitants of Asia Minor, including Greeks, Phyrigians, and Armenians, who went over to the Islamic religion and Turkic language, creating a frontier society. Armenian communities continued to flourish under relatively tolerant Ottoman rule for centuries, either as minority populations in urban areas or as exclusively Armenian towns in rural areas. In cities such as Istanbul and İzmir, Armenians played particularly important roles; an 1851 New York Times report, for instance, indicates that Armenians comprised nearly one quarter of the population of Istanbul at that time, with over 200,000 residents.\n", "The Ottoman Empire disintegrated, with much of its Levant territory awarded to various Allied powers as protectorates. The Turkish core in Anatolia was reorganised as the Republic of Turkey. The Ottoman Empire was to be partitioned by the Treaty of Sèvres of 1920. This treaty was never ratified by the Sultan and was rejected by the Turkish National Movement, leading to the victorious Turkish War of Independence and the much less stringent 1923 Treaty of Lausanne.\n", "Between 1918 and 1923, Turkish resistance movements led by Mustafa Kemal Atatürk forced the Greeks and Armenians out of Anatolia, while the Italians never established a presence. The Turkish revolutionaries also suppressed Kurdish attempts to become independent in the 1920s. After the Turkish resistance gained control over Anatolia, there was no hope of meeting the conditions of the Treaty of Sèvres.\n", "After the Imperial army suffered defeat in Asia Minor, Andronikos III saw Anatolia as a lost cause and began reorganizing the Byzantine fleet; as a result the Aegean remained an effective defense against Turkish incursions until Gallipoli was at last captured by the Turks in 1354. From then on, the Byzantine military engaged in small scale warfare against her weak Crusader opponents, mixing in diplomacy and subterfuge, often exploiting civil conflict amongst their Ottoman opponents. In the Peloponnese, territory continued to be re-conquered by the Byzantines against the weak crusaders until the mid 15th century, when the Byzantine enclave in Morea was finally conquered by the Ottomans.\n", "By their victory at Nicopolis, the Turks discouraged the formation of future European coalitions against them. They maintained their pressure on Constantinople, tightened their control over the Balkans, and became a greater threat to central Europe.\n", "In 1068 Alp Arslan and allied Turkmen tribes recaptured many Abbasid lands and even invaded Byzantine regions, pushing further into eastern and central Anatolia after a major victory at the Battle of Manzikert in 1071. The disintegration of the Seljuk dynasty, the first unified Turkic dynasty, resulted in the rise of subsequent, smaller, rival Turkic kingdoms such as the Danishmends, the Sultanate of Rum, and various Atabegs who contested the control of the region during the Crusades and incrementally expanded across Anatolia until the rise of the Ottoman Empire.\n", "Following the Ottoman defeat at World War I, the Ottoman capital Constantinople (modern Istanbul) and much of Anatolia were occupied by the Allies, who planned to share these lands between Armenia, France, Greece, Italy and the United Kingdom, leaving for the Turks the core piece of land in central Anatolia. In response, the leader of the Turkish nationalist movement, Mustafa Kemal Atatürk, established the headquarters of his resistance movement in Angora in 1920. After the Turkish War of Independence was won and the Treaty of Sèvres was superseded by the Treaty of Lausanne (1923), the Turkish nationalists replaced the Ottoman Empire with the Republic of Turkey on 29 October 1923. A few days earlier, Angora had officially replaced Constantinople as the new Turkish capital city, on 13 October 1923, and Republican officials declared that the city's name is Ankara.\n" ]
Is there any scientific truth to the idea that you shouldn't sit too close to the TV?
No. However focusing on one spot for a long period of time can cause eyestrain, however that is only temporary. You may find this interesting. _URL_0_
[ "The work of several researchers support the concept of television reality as a consequence of heavy viewing. According to Wyer and Budesheim's research, television messages or information, even when they are not necessarily considered truthful, can still be used in the process of constructing social judgments. Furthermore, indicted invalid information may still be used in subsequent audience's judgments.\n", "The theory has also received criticism for ignoring other issues such as the perceived realism of the televised content, which could be essential in explaining people's understanding of reality. Wilson, Martins, & Markse (2005) argue that attention to television might be more important to cultivating perceptions than only the amount of television viewing. In addition, C. R. Berger (2005) writes that because the theory ignores cognitive processes, such as attention or rational thinking style, it is less useful than desired.\n", "In this case, television forms a reality in the minds of the people, specifically, as an \"ideal\" world on where to live in. The people perceive it and then in turn form the notion that \"this is how life should be.\" This happens, when, in fact, it should not be, because the television is not the reality but life is the ultimate reality.\n", "The television, along with the remote control, is also argued to be conditioning humans into habitually repeating that which they enjoy without experiencing genuine surprise or even discomfort, a critique of the television similar to that of those made against information silos on social media sites today. In essence, this technological development led to \"egocasting\": a world in which people exert extreme amounts of control over what they watch and hear. As a result, users deliberately avoid content they disagree with in any form––ideas, sounds, or images. In turn, this siloing can drive political polarization and stoke tribalism. Plus, companies like TiVO analyze how TV show watchers use their remote and DVR capability to skip over programming, such as advertisements––a privacy concern users may lack awareness of as well.\n", "BULLET::::- \"So I repeat that while theoretically and technically television may be feasible, yet commercially and financially, I consider it an impossibility; a development of which we need not waste little time in dreaming.\" – 1926\n", "Activist Jerry Mander's book argues for the complete removal of television from our lives because of its ill effects. Mander gives the example of Tausk's \"Influencing machine\" as being a parallel for television: \"Doubtless you have noticed that this 'influencing machine' sounds an awful lot like television ... In any event, there is no question that television does what the schizophrenic fantasy says it does. It places in our minds images of reality which are outside our experience. The pictures come in the form of rays from a box. They cause changes in feeling and ... utter confusion as to what is real and what is not.\"\n", "It is Masthoff's contention that it is possible to develop an interactive, personalized form of television that would allow the viewer to engage in natural conversation and learn from these conversations.\n" ]
how is it that a vehicle can be good at towing but shit at carrying heavy loads?
A vehicle can have a strong engine, trans and drive train to pull a heavy tow. It needs a strong suspension system to carry a heavy load.
[ "The military also deploys tow trucks for recovery of stranded vehicles. In the US Army, a variant of the HEMTT truck is used for this purpose, the M984 wrecker. For recovery in combat situations while under fire, many armies with large vehicle fleets also deploy armoured recovery vehicles. These vehicles fulfill a similar role, but are resistant to heavy fire and capable of traversing rough terrain with their tracks, as well as towing vehicles beyond the weight limits of wheeled wreckers, such as tanks (many are based on tank designs for this very reason).\n", "BULLET::::- A tow truck is a vehicle used to transport motor vehicles to another location, generally a repair garage, or to recover vehicles which are no longer on a drivable surface. Vehicles are often towed in the case of breakdowns or collisions, or may be impounded for legal reasons. The tow truck was invented in 1916 by Ernest Holmes, Sr., of Chattanooga, Tennessee. He was a garage worker who was inspired to create the invention after he was forced to pull a car out of a creek using blocks, ropes, and six men. An improved design led him to manufacture wreckers.\n", "Hitch tow trucks are mostly sized for cars and light duty trucks. Larger versions, with a long, weighted body and heavier duty engines, transmissions, and tow hooks, may be used for towing of disabled buses, truck tractors, or large trucks. The artificial sizing and weighting must be designed to withstand the greater weight of the towed vehicle, which might otherwise tip the tow truck back.\n", "Towing capacity is a measure describing the upper limit to the weight of a trailer a vehicle can tow and may be expressed in pounds or kilograms. Some countries require that signs indicating the maximum trailer weight (and in some cases, length) be posted on trucks and buses close to the coupling device. Towing capacity may be lower as declared due to limitation imposed by the cooling system.\n", "A tow truck (also called a wrecker, a breakdown truck, recovery vehicle or a breakdown lorry) is a truck used to move disabled, improperly parked, impounded, or otherwise indisposed motor vehicles. This may involve recovering a vehicle damaged in an accident, returning one to a drivable surface in a mishap or inclement weather, or towing or transporting one via flatbed to a repair shop or other location.\n", "Towing may be as simple as a tractor pulling a tree stump. The most familiar form is the transport of disabled or otherwise indisposed vehicles by a tow truck or \"wrecker.\" Other familiar forms are the tractor-trailer combination, and cargo or leisure vehicles coupled via ball or pintle and gudgeon trailer-hitches to smaller trucks and cars. In the opposite extreme are extremely heavy duty tank recovery vehicles, and enormous ballast tractors involved in heavy hauling towing loads stretching into the millions of pounds.\n", "A heavy rescue vehicle is a type of specialty emergency medical services or firefighting apparatus. They are primarily designed to provide the specialized equipment necessary for technical rescue situations, as well as search and rescue within structure fires. They carry an array of special equipment such as the Jaws of life, wooden cribbing, generators, winches, hi-lift jacks, cranes, cutting torches, circular saws and other forms of heavy equipment unavailable on standard trucks. This capability differentiates them from traditional pumper trucks or ladder trucks designed primarily to carry firefighters and their entry gear as well as on-board water tanks, hoses and equipment for fire extinguishing and light rescue. Most heavy rescue vehicles lack on-board water tanks and pumping gear, owing to their specialized role, but some do carry on-board pumps in order to broaden their response capability.\n" ]
What is the chemical reaction that occurs when you put ice cream into soft drink?
No chemical reaction... the fat and sugars from the ice cream increases the surface tension on the bubbles, which means they don't burst. You see a bunch of bubbles because you've just introduced something with a bunch of nucleation sites.
[ "Ice cream (derived from earlier iced cream or cream ice) is a sweetened frozen food typically eaten as a snack or dessert. It may be made from dairy milk or cream, or soy, cashew, coconut or almond milk, and is flavored with a sweetener, either sugar or an alternative, and any spice, such as cocoa or vanilla. Colourings are usually added, in addition to stabilizers. The mixture is stirred to incorporate air spaces and cooled below the freezing point of water to prevent detectable ice crystals from forming. The result is a smooth, semi-solid foam that is solid at very low temperatures (below ). It becomes more malleable as its temperature increases.\n", "BULLET::::- Affogato is a cold drink, often served as dessert, consisting of a scoop of ice cream or gelato topped with an espresso shot. Often, the drinker is served the ice cream and espresso in separate cups, and will mix them at the table so as to prevent the ice cream from entirely melting before it can be consumed.\n", "Traditional ice cream is frozen egg custard with flavours added. Blumenthal whisks egg yolks with sugar until the sugar interacts with the proteins in the yolk, creating a network of proteins. The entire substance turns white, at which point flavouring can be added and cooked in. While stirring the mixture, Blumenthal cools it as fast as possible using liquid nitrogen.\n", "Originating in Australia and similar to the Mazagran, the minimal Ice Shot is a single shot of fresh espresso poured into an ordinary latté glass that has been filled with ice. The hot coffee, in melting some of the ice is diluted, re-freezing to a granita-like texture. The addition of a single scoop of ice-cream on top is a popular variant. No milk, sugar, extra flavouring or cream are involved.\n", "BULLET::::- The Ice cream mix is defined as the pasteurized mix of cream, milk and other milk products that are not yet frozen. It may contain eggs, artificial or non-artificial flavours, cocoa or chocolate syrup, a food color, an agent that adjusts the pH level in the mix, salt, a stabilizing agent that doesn’t exceed 0.5% of the ice cream mix, a sequestering agent which preserves the food colour, edible casein that doesn’t exceed 1% of the mix, propylene glycol mono fatty acids in an amount that will not exceed 0.35% of the ice cream mix and sorbitan tristearate in an amount that will not exceed 0.035% of the mix. The ice cream mix may not include less than 36% solid components.\n", "The dessert is commonly made by taking a scoop of ice cream frozen well below the temperature at which ice cream is generally kept, possibly coating it in raw egg, rolling it in cornflakes or cookie crumbs, and briefly deep frying it. The extremely low temperature of the ice cream prevents it from melting while being fried. It may be sprinkled with cinnamon and sugar and a touch of peppermint, though whipped cream or honey may be used as well.\n", "An ice cream maker has to simultaneously freeze the mixture while churning it so as to aerate the mixture and keep the ice crystals small (less than 50 μm). As a result, most ice creams are ready to consume immediately. However, those containing alcohol must often be chilled further to attain a firm consistency.\n" ]
Why do we discuss Ancient Greece in terms of city-states?
Many Greek city-states had slightly different setups, but their basic structure was the same: a single *astu* (city-center) with administrative control over a *chora* (territory). Athens controlled a large *chora* (Attica) as the result of *synoekismos,* or synoecism as the Brits would say. This is the process by which, usually early on in the Iron Age, various smaller population centers came together (literally "housed" together) to form a larger political unit: a *polis*, comprised of *astu* and *chora.* Why one center became the *astu* vs another was usually dependent on population, military might, religious importance, or the like. In Attica, the smaller population centers were organized into districts ("demes"). Some examples are the Piraeus (the harbor), or Sunion (a temple center in the south of Attica). These were not "cities" in that they had no independent authority in the business of Athens the polis. You should think of them as neighborhoods within a modern city, with some localized civic or religious structures.
[ "Athens was chosen as the Greek capital for historical and sentimental reasons, not because it was a large city. At the time, it was a town consisting of only 400 houses at the foot of the Acropolis. A modern city plan was laid out, and public buildings erected. The finest legacy of this period are the buildings of the University of Athens (1837, under the name Othonian University), the Athens Polytechnic University (1837, under the name Royal School of Arts), the National Gardens of Athens (1840), the National Library of Greece (1842), the Old Royal Palace (now the Greek Parliament Building, 1843), and the Old Parliament Building (1858). Schools and hospitals were established all over the (still small) Greek dominion, Due to the negative feelings of the Greek people toward non-Greek rule, historical attention to this aspect of his reign has been neglected.\n", "Greece is considered the cradle of Western civilisation, being the birthplace of democracy, Western philosophy, Western literature, historiography, political science, major scientific and mathematical principles, Western drama and notably the Olympic Games. From the eighth century BC, the Greeks were organised into various independent city-states, known as \"poleis\" (singular \"polis\"), which spanned the entire Mediterranean region and the Black Sea. Philip of Macedon united most of the Greek mainland in the fourth century BC, with his son Alexander the Great rapidly conquering much of the ancient world, from the eastern Mediterranean to India. Greece was annexed by Rome in the second century BC, becoming an integral part of the Roman Empire and its successor, the Byzantine Empire, in which Greek language and culture were dominant. Rooted in the first century A.D., the Greek Orthodox Church helped shape modern Greek identity and transmitted Greek traditions to the wider Orthodox World. Falling under Ottoman dominion in the mid-15th century, the modern nation state of Greece emerged in 1830 following a war of independence. Greece's rich historical legacy is reflected by its 18 UNESCO World Heritage Sites.\n", "Classical Athens was a powerful city-state that emerged in conjunction with the seagoing development of the port of Piraeus. A center for the arts, learning and philosophy, home of Plato's Academy and Aristotle's Lyceum, it is widely referred to as the cradle of Western civilization and the birthplace of democracy, largely because of its cultural and political impact on the European continent, and in particular the Romans. In modern times, Athens is a large cosmopolitan metropolis and central to economic, financial, industrial, maritime, political and cultural life in Greece.\n", "The emergence of city-states (\"poleis\") in ancient Greece marks the beginning of classical antiquity. The two most important Greek cities, the Ionian-democratic Athens and the Dorian-aristocratic Sparta, led the successful defense of Greece against the invading Persians from the east, but then clashed against each other for supremacy in the Peloponnesian War. The Kingdom of Macedon took advantage of the following instability and established a single rule over Greece. Desire to form a universal monarchy brought Alexander the Great to annex the entire Persian Empire and begin a hellenization of the Macedonian possessions. At his death in 323 BC, his reign was divided between his successors and several hellenistic kingdoms were formed.\n", "Ancient Greece consisted of several hundred relatively independent city-states (\"poleis\"). This was a situation unlike that in most other contemporary societies, which were either tribal or kingdoms ruling over relatively large territories. Undoubtedly the geography of Greece—divided and sub-divided by hills, mountains, and rivers—contributed to the fragmentary nature of ancient Greece. On the one hand, the ancient Greeks had no doubt that they were \"one people\"; they had the same religion, same basic culture, and same language. Furthermore, the Greeks were very aware of their tribal origins; Herodotus was able to extensively categorise the city-states by tribe. Yet, although these higher-level relationships existed, they seem to have rarely had a major role in Greek politics. The independence of the \"poleis\" was fiercely defended; unification was something rarely contemplated by the ancient Greeks. Even when, during the second Persian invasion of Greece, a group of city-states allied themselves to defend Greece, the vast majority of \"poleis\" remained neutral, and after the Persian defeat, the allies quickly returned to infighting.\n", "The Greeks and the Romans left a legacy in Europe which is evident in European languages, thought, visual arts and law. Ancient Greece was a collection of city-states, out of which the original form of democracy developed. Athens was the most powerful and developed city, and a cradle of learning from the time of Pericles. Citizens' forums debated and legislated policy of the state, and from here arose some of the most notable classical philosophers, such as Socrates, Plato, and Aristotle, the last of whom taught Alexander the Great.\n", "The territory of Greece is mountainous, and as a result, ancient Greece consisted of many smaller regions each with its own dialect, cultural peculiarities, and identity. Regionalism and regional conflicts were a prominent feature of ancient Greece. Cities tended to be located in valleys between mountains, or on coastal plains, and dominated a certain area around them.\n" ]
what would happen if there was a second big bang somewhere else outside of our own expanding universe?
If it was outside of our universe it would also be beyond everything that we are capable of observing or understanding, so we would have no way of knowing that it even happened.
[ "The Big Bang itself had been proposed in 1931, long before this period, by Georges Lemaître, a Belgian physicist, who suggested that the evident expansion of the Universe in time required that the Universe, if contracted backwards in time, would continue to do so until it could contract no further. This would bring all the mass of the Universe to a single point, a \"primeval atom\", to a state before which time and space did not exist. Hoyle is credited with coining the term \"Big Bang\" during a 1949 BBC radio broadcast, saying that Lemaître's theory was \"based on the hypothesis that all the matter in the universe was created in one big bang at a particular time in the remote past.\" It is popularly reported that Hoyle intended this to be pejorative, but Hoyle explicitly denied this and said it was just a striking image meant to highlight the difference between the two models. Lemaître's model was needed to explain the existence of deuterium and nuclides between helium and carbon, as well as the fundamentally high amount of helium present, not only in stars but also in interstellar space. As it happened, both Lemaître and Hoyle's models of nucleosynthesis would be needed to explain the elemental abundances in the universe.\n", "The Designer Universe theory of John Gribbin suggests that the universe could have been made deliberately by an advanced civilization in another part of the Multiverse, and that this civilization may have been responsible for causing the Big Bang.\n", "The Big Bang is not an explosion of matter moving outward to fill an empty universe. Instead, space itself expands with time everywhere and increases the physical distance between two comoving points. In other words, the Big Bang is not an explosion \"in space\", but rather an expansion \"of space\". Because the FLRW metric assumes a uniform distribution of mass and energy, it applies to our universe only on large scales—local concentrations of matter such as our galaxy are gravitationally bound and as such do not experience the large-scale expansion of space.\n", "A more specific theory called \"Big Bounce\" proposes that the universe could collapse to the state where it began and then initiate another Big Bang, so in this way the universe would last forever, but would pass through phases of expansion (Big Bang) and contraction (Big Crunch). \n", "Most cosmologists today believe that the universe expanded from a singularity approximately 13.8 billion years ago in a 'smeared-out singularity' called the Big Bang, meaning that space itself came into being at the moment of the big bang and has expanded ever since, creating and carrying the galaxies with it.\n", "This scenario allows the Big Bang to occur immediately after the Big Crunch of a preceding universe. If this happens repeatedly, it creates a cyclic model, which is also known as an oscillatory universe. The universe could then consist of an infinite sequence of finite universes, with each finite universe ending with a Big Crunch that is also the Big Bang of the next universe. Theoretically, the cyclic universe could not be reconciled with the second law of thermodynamics: entropy would build up from oscillation to oscillation and cause heat death. Current evidence also indicates the universe is not closed. This has caused cosmologists to abandon the oscillating universe model. A somewhat similar idea is embraced by the cyclic model, but this idea evades heat death because of an expansion of the branes that dilutes entropy accumulated in the previous cycle.\n", "Subsequent modelling of the universe explored the possibility that the cosmological constant, introduced by Einstein in his 1917 paper, may result in an expanding universe, depending on its value. Thus the Big Bang model was proposed by the Belgian priest Georges Lemaître in 1927 which was subsequently corroborated by Edwin Hubble's discovery of the redshift in 1929 and later by the discovery of the cosmic microwave background radiation by Arno Penzias and Robert Woodrow Wilson in 1964. These findings were a first step to rule out some of many alternative cosmologies.\n" ]
the four tigers of asia
You could post this question in /r/AskSocialScience if it hasn't been asked there already.
[ "\"Widely known as one of the largest predators, tigers are ranged both in Russia and China, where this species is even considered sacred. Moreover, the form of Chinese character 王, which means “a king”, reminds us of stripe pattern on the tiger’s forehead. White tigers, in particular, adapt easily to any environment if there is enough free space, being as well the largest representatives of their breed. In Chinese mythology the white tiger symbolizes courage and strong spirit that protects one from external threats. \n", "The plain tiger is found across the entirety of Africa, where the predominant subspecies is \"D. c. alcippus\". Its range extends across the majority of Asia throughout Indian subcontinent, as well as many south Pacific islands. The plain tiger is even present in parts of Australia. \"D. c. chrysippus\" is most common throughout Asia and in some select regions in Africa, while \"D. c. orientis\" is present in more tropical African regions as well as some African islands, including Madagascar and the Seychelles.\n", "The Siberian tiger (\"Panthera tigris tigris\") is a tiger population in the Russian Far East and Northeast China, and possibly North Korea. It once ranged throughout the Korean Peninsula, north China, Russian Far East, and eastern Mongolia. Today, this population inhabits mainly the Sikhote Alin mountain region in southwest Primorye Province in the Russian Far East. In 2005, there were 331–393 adult and subadult Siberian tigers in this region, with a breeding adult population of about 250 individuals. The population had been stable for more than a decade due to intensive conservation efforts, but partial surveys conducted after 2005 indicate that the Russian tiger population was declining. An initial census held in 2015 indicated that the Siberian tiger population had increased to 480–540 individuals in the Russian Far East, including 100 cubs. This was followed up by a more detailed census which revealed there was a total population of 562 wild Siberian tigers in Russia. As of 2014, about 35 individuals were estimated to range in the international border area between Russia and China.\n", "The validity of several tiger subspecies in continental Asia was questioned in 1999. Morphologically, tigers from different regions vary little, and gene flow between populations in those regions is considered to have been possible during the Pleistocene. Therefore, it was proposed to recognise only two subspecies as valid, namely \"P. t. tigris\" in mainland Asia, and \"P. t. sondaica\" in the Greater Sunda Islands and possibly in Sundaland.\n", "In East Asia, the tiger inhabits Korean pine and temperate broadleaf and mixed forests in the Amur-Ussuri region of Primorsky Krai and Khabarovsk Krai in far eastern Siberia. Riparian forests are important habitats for both ungulates and tigers as they provide food and water, and serve as dispersal corridors.\n", "The tiger once ranged widely from the Eastern Anatolia Region in the west to the Amur River basin, and in the south from the foothills of the Himalayas to Bali in the Sunda islands. Since the early 20th century, tiger populations have lost at least 93% of their historic range and have been extirpated in Western and Central Asia, from the islands of Java and Bali, and in large areas of Southeast and South Asia and China. Today's tiger range is fragmented, stretching from Siberian temperate forests to subtropical and tropical forests on the Indian subcontinent and Sumatra. The tiger is listed as Endangered on the IUCN Red List since 1986. As of 2015, the global wild tiger population was estimated to number between 3,062 and 3,948 mature individuals, down from around 100,000 at the start of the 20th century, with most remaining populations occurring in small pockets isolated from each other. Major reasons for population decline include habitat destruction, habitat fragmentation and poaching. This, coupled with the fact that it lives in some of the more densely populated places on Earth, has caused significant conflicts with humans.\n", "For the naming of the team, South China tigers are considered as the most distinctive of all tiger subspecies. The population once numbered more than 4,000 in the wild, distributed from Hunan, Jiangxi in the north to as far south as Hong Kong.\n" ]
why is there so much nudity in classical paintings?
So with the spread of humanism ushered in by the beginning of the renaissance, people started to take a keen interest in humanity. They started thinking that people were incredible and the individual was exceptionally important. As a result, we start seeing changes in a number of artistic modes of expression. To answer your question though, painters and sculptors wanted to idealize the human form which took on a whole new significance in renaissance art. The natural human figure presented an expression of beauty, perfection, and humanity. TL;DR - Round about 1400 or so, people started thinking of humans as hot shit. As a result they celebrated the ideal human form - a nude figure.
[ "In comparison in the material aspect nudity is considered an art. This view is supported by Sri Aurobindo in his book \"The Renaissance in India\". He says about Hinduism in the book – \"Its spiritual extremism could not prevent it from fathoming through a long era the life of the senses and its enjoyments, and there too it sought the utmost richness of sensuous detail and the depths and intensities of sensual experience. Yet it is notable that this pursuit of the most opposite extremes never resulted in disorder…\"\n", "Partly out of interest in the natural world and partly out of interest in the classical cultures of ancient Greece and Rome, portraits—both painted and sculpted—were given an important role in Renaissance society and valued as objects, and as depictions of earthly success and status. Painting in general reached a new level of balance, harmony, and insight, and the greatest artists (Leonardo, Michelangelo, and Raphael) were considered \"geniuses\", rising far above the tradesman status to valued servants of the court and the church.\n", "Kenneth Clark noted that sexuality was part of the attraction to the nude as a subject of art, stating \"no nude, however abstract, should fail to arouse in the spectator some vestige of erotic feeling, even though it be only the faintest shadow—and if it does not do so it is bad art and false morals.\" According to Clark, the explicit temple sculptures of tenth-century India \"are great works of art because their eroticism is part of their whole philosophy.\" Great art can contain significant sexual content without being obscene. However sexually explicit works of fine art produced in Europe before the modern era, such as Gustav Courbet's \"L'Origine du monde\", were not intended for public display. The judgement of whether a particular work is artistic or pornographic is ultimately subjective and has changed through history and from one culture to another. Some individuals judge any public display of the unclothed body to be unacceptable, while others may find artistic merit in explicitly sexual images. Public reviews of art may or may not address the issue.\n", "On the other hand, it has been suggested that Leonardo da Vinci may have had Pollaiuolo partly in mind when he wrote that artists should not:make their nudes wooden and without grace, so that they seem to look like a sack of nuts rather than the surface of a human being, or indeed a bundle of radishes rather than muscular nudes\n", "In Greece and Rome, wall painting was not considered as high art. The most prestigious form of art besides sculpture was panel painting, i.e. tempera or encaustic painting on wooden panels. Unfortunately, since wood is a perishable material, only a very few examples of such paintings have survived, namely the Severan Tondo from circa 200 AD, a very routine official portrait from some provincial government office, and the well-known Fayum mummy portraits, all from Roman Egypt, and almost certainly not of the highest contemporary quality. The portraits were attached to burial mummies at the face, from which almost all have now been detached. They usually depict a single person, showing the head, or head and upper chest, viewed frontally. The background is always monochrome, sometimes with decorative elements. In terms of artistic tradition, the images clearly derive more from Greco-Roman traditions than Egyptian ones. They are remarkably realistic, though variable in artistic quality, and may indicate the similar art which was widespread elsewhere but did not survive. A few portraits painted on glass and medals from the later empire have survived, as have coin portraits, some of which are considered very realistic as well. Pliny the Younger complained of the declining state of Roman portrait art, \"The painting of portraits which used to transmit through the ages the accurate likenesses of people, has entirely gone out…Indolence has destroyed the arts.\"\n", "Murals were a common form of wall decoration in ancient Rome. The earliest Christian mural paintings come from the catacombs of Rome. They include many representations of Christ as \"the Good Shepherd\", generally as a standardised image of a young, beardless man with a sheep on his shoulders. Other popular subjects include the \"Madonna and Child\", Jonah being thrown into the sea, the three young men in the furnace and the \"Last Supper\". In one remarkable mural, in the Catacomb of the Aurelii, is the earliest image of Jesus, as he came to be commonly depicted, as a bearded, Jewish man in long robes. In this particular image he is preaching, not to a group of people but to a flock of sheep and goats, representing the faithful and the wayward.\n", "As a result of the Renaissance, in many European societies artists were strongly influenced by classical Greek styles and culture. As a result, images of nude and semi-nude subjects in many forms proliferated in art and sculpture.\n" ]
why do people pass out/feel like they're about to, if the suddenly go from very warm water to very cold water or vice versa?
It's because of the mammalian diving reflex. We experience bradycardia (slow heart rate) and vasoconstriction. Essentially all of our oxygenated blood is shunted to vital areas (Brain, heart, and lungs)only. This conserves oxygen allowing us to endure the cold water for a longer period of time for the best possible chance at survival. Unfortunately in non life threatening situations it can make is feel like we're going to pass out. This "hack" is actually used as a first line treatment by physicians when a patient has a rapid heart rate.
[ "Cold shock response is the physiological response of organisms to sudden cold, especially cold water, and is a common cause of death from immersion in very cold water, such as by falling through thin ice. The immediate shock of the cold causes involuntary inhalation, which if underwater can result in drowning. The cold water can also cause heart attack due to vasoconstriction; the heart has to work harder to pump the same volume of blood throughout the body, and for people with heart disease, this additional workload can cause the heart to go into arrest. A person who survives the initial minute after falling into cold water can survive for at least thirty minutes provided they do not drown. The ability to stay afloat declines substantially after about ten minutes as the chilled muscles lose strength and co-ordination.\n", "Cold shock response is the physiological response of organisms to sudden cold, especially cold water, and is a common cause of death from immersion in very cold water, such as by falling through thin ice. The immediate shock of the cold causes involuntary inhalation, which if underwater can result in drowning. The cold water can also cause heart attack due to vasoconstriction; the heart has to work harder to pump the same volume of blood throughout the body, and for people with heart disease, this additional workload can cause the heart to go into arrest. A person who survives the initial minute of trauma after falling into icy water can survive for at least thirty minutes provided they don't drown. However, the ability to perform useful work like staying afloat declines substantially after ten minutes as the body protectively cuts off blood flow to \"non-essential\" muscles.\n", "In humans, cold shock response is perhaps the most common cause of death from immersion in very cold water, such as by falling through thin ice. The immediate shock of the cold causes involuntary inhalation, which if underwater can result in drowning. The cold water can also cause heart attack due to vasoconstriction; the heart has to work harder to pump the same volume of blood throughout the body. For people with existing cardiovascular disease, the additional workload can result in cardiac arrest. Inhalation of water (and thus drowning) may result from hyperventilation. Some people are much better able to survive swimming in very cold water due to body or mental conditioning.\n", "Winter swimming can be dangerous to people who are not used to swimming in very cold water. After submersion in cold water the cold shock response will occur, causing an uncontrollable gasp for air. This is followed by hyperventilation, a longer period of more rapid breathing. The gasp for air can cause a person to ingest water, which leads to drowning. As blood in the limbs is cooled and returns to the heart, this can cause fibrillation and consequently cardiac arrest. The cold shock response and cardiac arrest are the most common causes of death related to cold water immersion.\n", "Cold shock response is the initial reaction to immersion in cold water. It generally starts with a gasp reflex in response to sudden and rapid chilling of the skin, and if the head is immersed there is a risk of inhaling water and drowning. This is followed by a reflexive hyperventilation, with a risk of panic and fainting if not controlled. Cold induced vasoconstriction causes the heart to work harder and the additional work can overload a weak heart, with a possible consequence of cardiac arrest. Cold incapacitation is the next stage, and generally occurs within 5 to 15 minutes in cold water. Blood flow to the extremities is reduced by vasoconstriction as the body attempts to reduce heat loss from the vital organs of the core. This accelerates the cooling of the periphery, and reduces the functionality of the muscles and nerves. The duration of exposure to produce hypothermia varies with health, body mass and water temperature. It generally takes in the order of 30 minutes for an unprotected person in water to become hypothermic.\n", "Hypothermia from exposure to cold water is not as sudden as is often believed. A person who survives the initial minute of trauma (after falling into icy water), can survive for at least thirty minutes provided they don't drown. However, the ability to perform useful work (for example to save oneself) declines substantially after ten minutes (as the body protectively cuts off blood flow to \"non-essential\" muscles).\n", "The actual cause of death in cold water is usually the bodily reactions to heat loss and to freezing water, rather than hypothermia (loss of core temperature) itself. For example, plunged into freezing seas, around 20% of victims die within two minutes from cold shock (uncontrolled rapid breathing, and gasping, causing water inhalation, massive increase in blood pressure and cardiac strain leading to cardiac arrest, and panic); another 50% die within 15–30 minutes from cold incapacitation (inability to use or control limbs and hands for swimming or gripping, as the body \"protectively\" shuts down the peripheral muscles of the limbs to protect its core). Exhaustion and unconsciousness cause drowning, claiming the rest within a similar time.\n" ]
with woman's equality in military combat roles being normalized, why is it still absurd that woman play with men in the nfl and other major league sports?
Because physically a woman in the NFL would not be able to stop most of the males. Now, if you have some sort of 6'4 275 she-hulk that would work.
[ "Finally, there is the argument that by not incorporating women into combat, the American government is failing to tap into another source of soldiers for military combat operations. This argument claims that the government is creating a military that treats women as second-class citizens and not equals of men.\n", "Women in the military face a similar problem. Recent wars in Iraq and Afghanistan have allowed women combat roles. However, in order for women in the military to be accepted and considered successful, they feel they must become \"one of the guys.\" Otherwise they face sexual and gender based ridicule that, in some cases, led to women ending their military careers. Feminist theorist Cynthia Enloe argues that the institution of the military is not comparable to those of education or business because of its inherent violent and hyper-masculine characteristics. She states that this environment is so harmful for women that they can never fully assimilate.\n", "Women serving in the U.S. military in the past have often seen combat despite the Combat Exclusion Policy. Due to a shortage of troops, women were temporarily attached to direct combat units slipping in through a bureaucratic loophole. Although they were not supposed to be in positions that engaged in direct combat, and were ineligible for combat pay, thousands of women have engaged the enemy directly in Operations Iraqi and Enduring Freedom.\n", "The Army and Marine Corps precluded the use of women in combat as a matter of established policy, and both the Navy and the Air Force restricted women's participation in combat. Even the president, who had originally suggested that women be included, expressed his intent to continue the current military policy excluding women from combat. Since the purpose of registration was to prepare for a draft of combat troops, and since women are excluded from combat, Congress concluded that they would not be needed in the event of a draft, and therefore decided funds should not be used to register them. As one Senator said, “It has been suggested that all women be registered, but only a handful actually be inducted in an emergency. The Committee finds this a confused and ultimately unsatisfactory solution.\" As the Senate Committee recognized a year before, \"training would be needlessly burdened by women recruits who could not be used in combat.\" All in all, the proponents of the current MSSA advocated not using government funds to register people who were excluded from the job anyway. The main point of those who favored the registration of females was that females were in favor of it because of gender equality principles; women, as full citizens, ought to have the same civic duties and responsibilities as men.\n", "Brian Mitchell, in his article \"Women Make Poor Soldiers\" (excerpted from his 1989 book \"\"Weak Link: The Feminization of the American Military\"\"), expressed concern that placing women in combat lowers unit cohesion, either due to sexual relationships taking priority over group loyalty, or because men would feel obliged to be more protective of women than other men. Mitchell's view was harshly criticized in a New York Times review, which stated the book was \"spoiled by intemperate allegations and a supercilious tone\" and lacked sourcing for statements.\n", "In the United States, women are seen as 'ill-equipped' to participate in sports, and their involvement was viewed as unfeminine and undesirable. The reasons why women experience less academic advantage from sports than men do focus on the clash between expectations for women, athletes, and the stigma for female athletes who are seen to be unfeminine.\n", "As of 2013, the only sports that men, but not women play professionally in the United States are football, baseball, and Ultimate Frisbee. Although basketball, soccer and hockey have female sports leagues, they are far behind in terms of exposure and funding compared to the men's teams.\n" ]
what is the body trying to accomplish when we dry heave?
Your body is attempting to vomit but not succeeding. That's pretty much it.
[ "Other than treating, curing or remedying the underlying cause of emaciation, it as a symptom is treated by regaining the weight and restoring the tissues. This is done through renourishment, or reintroducing nourishing liquids and foods to the body while increasing the intake of food energy. The process, usually begun in an individual deprived of food for a period of time, must be done slowly to avoid complications such as regurgitation and vomiting. It begins with spoonfuls of water and salted broth, advancing to increased amounts of clear liquids including broth, tea and fruit juices. This soon is advanced to full liquid diet such as milk (if no lactose intolerance is present) and cream-based soups. Once solid food is introduced, an emaciated individual is usually given up to eight small meals per day, at two-hour intervals. Meals may consist of a small milkshake to minor portions of meat with a starchy side item. For the purposes of weight gain and tissue rebuilding, the diet will be focused on proteins, fats and carbohydrates that are rich in vitamins and minerals, and relatively high in energy. Oily foods and high-fiber foods like grains and certain vegetables are discouraged because they are difficult to digest, and filling while lower in energy. Treatment of emaciation also includes much sleep, rest and relaxation, and counseling.\n", "A dry enema is an alternative technique for cleansing the human rectum either for reasons of health, or for sexual hygiene. It is accomplished by squirting a small amount of sterile lubricant into the rectum, resulting in a bowel movement more quickly and with less violence than can be achieved by an oral laxative.\n", "Drying (dehydrating) is one of the oldest and easiest methods of food preservation. Dehydration is the process of removing water or moisture from a food product by heating at right temperature as well as containing air movement and dry air to absorb and carry the released moisture away. Reducing the moisture content of food prevents the growth of microorganisms such as bacteria, yeast and molds and slows down enzymatic reactions that take place within food. The combination of these events helps to prevent spoilage in dried food.\n", "The bulk of the water intake of \"Hemilepistus reaumuri\" is by taking up water vapour from saturated air and by eating damp sand. Water loss is minimised by the rectal epithelium, which absorbs water, ensuring that the faeces is drier than the food the animal consumed. Evaporation of water through the permeable exoskeleton may, however, provide a valuable cooling effect.\n", "A rudimentary form of \"dry\" enema is the use of a non-medicated glycerin suppository. However, due to the relative hardiness of the suppository - necessary for its insertion into the human body - before the glycerin can act, it must be melted by the heat of the body, and hence it does not take effect for up to an hour. Often the hygroscopic glycerin irritates the sensitive membranes of the rectum resulting in forceful expulsion of the suppository without any laxative effects.\n", "Involuntary movements may make feeding difficult or messy and may excessively prolong mealtimes. It may be easier to finger feed than use utensils (e.g., spoon or fork). For liquids, it is often easier to drink from a closed container with a straw than from an open cup. Caregivers may need to provide foods or liquids so that self-feeding is possible, or they may need to feed the person with A–T. In general, meals should be completed within approximately 30 minutes. Longer meals may be stressful, interfere with other daily activities, and limit the intake of necessary liquids and nutrients.\n", "Some \"Phyllomedusa\" species produce a waxy secretion that reduces the evaporative water loss of their bodies. If they begin to dry out, they move their limbs over their backs, where the secretory glands are, and spread the lipid secretion over their entire skin.\n" ]
Why were there only two American Aces in the Vietnam War?
Remember that becoming an ace requires downing five enemy aircraft, and to down enemy aircraft the enemy needs to have some in the first place. & #x200B; The VPAF had only started to receive jet fighters in 1964, and their aircraft were generally less numerous and less capable than what the Americans were fielding. Because the aims of the VPAF were exclusively defensive in nature and their supply of both manpower and materiel was very limited, they would deploy their aircraft very conservatively. They generally would only sortie their fighters when the situation was very much in their favor, and tactics were designed around ambushing strike formations, with forcing the attackers to drop their ordnance early and abort the mission being just as effective as downing an American aircraft. A more typical tactic involved having fighters appear away from a strike package to draw away escorts and then having another group pop in under the radar and make a pass on the strike aircraft themselves. While the typical strike aircraft (F-105) was nominally faster than the MiG-17, it could only outrun the MiG-17 once it had jettisoned its payload and thus ruined the mission. The MiGs, on the other hand, didn't exactly stay around to fight - all they had to do was ruin the mission, so a typical "attack" could consist of one pass on the strike aircraft followed by a rapid withdrawal by both the attacking and decoy aircraft. Because of that, opportunities for American fighters to actually engage the VPAF were few and far between. Making this worse, particularly early on, was the tendency of American forces to rotate pilots through the theater to spread experience and the generally poor air-to-air combat training provided early in the war. So while encountering an enemy fighter was already a rare thing for American pilots in Vietnam, the rotation of aircrews meant that it would be rare to stay in the theater long enough to encounter enough VPAF fighters to even have the potential to make ace, and poor air-to-air combat training meant that you were less likely to be successful at downing an enemy fighter before it was able to disengage and escape.
[ "The Vietnam War saw a move away from cannon fire to air-to-air missiles. Although US forces maintained air supremacy throughout the war, there were still occasional dogfights and US and North Vietnamese aces. The North Vietnamese side claimed the Vietnam People's Air Force had 17 aces throughout the war, including Nguyen Van Coc, who is also the top ace of Vietnam War with 9 kills: seven acknowledged by the United States Air Force.\n", "He became a pilot in the Marine Corps Reserve, but resigned to join the American Volunteer Group, better known as the Flying Tigers, to fight the Japanese prior to the United States entry into World War II. A member of the 3rd Pursuit Squadron (the \"Hell's Angels\"), he is credited with 10 victories, making him a double ace. By the end of the war, he had been promoted to lieutenant colonel.\n", "During World War II, the 17th Pursuit Squadron participated in the defense of the Philippines flying the Curtiss P-40 Warhawk and garnering the first American Ace of World War II. Wiped out during the Battle of the Philippines, some of its squadron members endured the Bataan Death March. Reactivated during the Vietnam War, the squadron went on to fly Republic F-105F Thunderchief Wild Weasel aircraft, and in Operation Desert Storm flying the General Dynamics F-16C Fighting Falcon.\n", "From April 1965 to November 1968, in 268 air battles conducted over North Vietnam, VPAF claimed to have shot down 244 US or ARVN's aircraft, and they lost 85 MiGs. During the war, 13 VPAF's flying aces attained their status while flying the MiG-21 (cf. three in the MiG-17).\n", "On 10 May 1972, Lieutenant Randy \"Duke\" Cunningham and Lieutenant (junior grade) William P. Driscoll flying an F-4J, call sign \"Showtime 100\", shot down three MiG-17s to become the first American flying aces of the war. Their fifth victory was believed at the time to be over a mysterious North Vietnamese ace, Colonel Nguyen Toon, now considered mythical. On the return flight, the Phantom was damaged by an enemy surface-to-air missile. To avoid being captured, Cunningham and Driscoll flew their burning aircraft using only the rudder and afterburner (the damage to the aircraft rendered conventional control nearly impossible), until they could eject over water.\n", "On October 24, 1944, in the initial phase of the Battle of Leyte Gulf, in the Philippines, he became the only American airman to achieve \"ace in a day\" status twice. McCampbell and his wingman attacked a Japanese force of 60 aircraft. McCampbell shot down nine, 7 Zeros and 2 Oscars, setting a U.S. single mission aerial combat record. During this same action, his wingman downed another six Japanese warplanes. When he landed his Grumman F6F Hellcat aboard USS \"Langley\" (the flight deck of \"Essex\" wasn't clear), his six machine guns had just two rounds remaining, and his airplane had to be manually released from the arrestor wire due to complete fuel exhaustion. Commander McCampbell received the Medal of Honor for both actions, becoming the only Fast Carrier Task Force pilot to be so honored.\n", "Various sources claim between 43 and 60 pilots from the Soviet Union attained ace status in the war. Most sources claim around 50 pilots attained ace status during the Korean War, of whom many are very controversial. Research by the USAF named 52 pilots who may have had legitimate claim to the title. Little is known of some of the pilots and their combined tally is incompatible with the number of aircraft the USAF claims to have lost in the war. Subsequent independent sources generally agree the number of aces claimed was around 52, but 15 names differ among the lists, particularly lower-scoring pilots. The number of victories for virtually all of the ace pilots is subject to dispute. Listed are names of 67 Soviet pilots attributed as aces in various sources. Of these, the ace status of 30 are in question among historians.\n" ]
Why did the brain of most animals evolve in the head and not in the torso?
It would be more protected, yes, but we keep our seeing, hearing, smelling, and tasting organs in our heads. If our brains were in our torsos, our senses of sight, smell, and hearing would be much less efficient (because the nerve signals would have to travel farther along pathways), and this would make us less able to survive. It's better to see a predator and run away before being attacked than it is to be attacked at all, even if your brain is better protected.
[ "Human brain – central organ of the nervous system located in the head of a human being, protected by the skull. It has the same general structure as the brains of other mammals, but with a more developed cerebral cortex than any other, leading to the evolutionary success of widespread dominance of the human species across the planet.\n", "In animals, the brain, or \"encephalon\" (Greek for \"in the head\"), is the control center of the central nervous system, responsible for thought. In most animals, the brain is located in the head, protected by the skull and close to the primary sensory apparatus of vision, hearing, equilibrioception, taste and olfaction. While all vertebrates have a brain, most invertebrates have either a centralized brain or collections of individual ganglia. Primitive animals such as sponges do not have a brain at all. Brains can be extremely complex. For example, the human brain contains around 86 billion neurons, each linked to as many as 10,000 others.\n", "The size and shape of the brain and the surrounding vault remain quite plastic as the brain grows in childhood. In several ancient societies, head shape was altered for aesthetic or religious reasons by binding cloth or boards tightly around the head during infancy. It is not known whether such artificial cranial deformation has an effect in brain power.\n", "All vertebrate brains share a common underlying form, which appears most clearly during early stages of embryonic development. In its earliest form, the brain appears as three swellings at the front end of the neural tube; these swellings eventually become the forebrain, midbrain, and hindbrain (the prosencephalon, mesencephalon, and rhombencephalon, respectively). At the earliest stages of brain development, the three areas are roughly equal in size. In many classes of vertebrates, such as fish and amphibians, the three parts remain similar in size in the adult, but in mammals the forebrain becomes much larger than the other parts, and the midbrain becomes very small.\n", "The evolution of a larger brain created a problem for early humans, however. A larger brain requires a larger skull, and thus requires the female to have a wider birth canal for the newborn's larger skull to pass through. But if the female's birth canal grew too wide, her pelvis would be so wide that she would lose the ability to run, which was a necessary skill 2 million years ago.\n", "When comparing the fossils with those of modern humans, the main difference is the elongated shape of the fossil braincase. According to the researchers, this indicates that brain shape, and possibly brain functions, evolved within the \"Homo sapiens\" lineage and relatively recently. Evolutionary changes in brain shape are likely to be associated with genetic changes of the brain's organization, interconnection and development and may reflect adaptive changes in the way the brain functions. Such changes may have caused the human brain to become rounder and two regions in the back of the brain to become enlarged over thousands of years of evolution.\n", "Heads develop in animals by an evolutionary trend known as cephalization. In bilaterally symmetrical animals, nervous tissues concentrate at the anterior region, forming structures responsible for information processing. Through biological evolution, sense organs and feeding structures also concentrate into the anterior region; these collectively form the head.\n" ]
why can some animals give birth without help when humans can't?
The only true assets of humanity are its intelligence and its social/communication skills. Those two assets require one big ass brain, proportionally. That giant brain is what makes our birth so difficult and also why we have such high mortality rates surrounding it. The other factor that makes birth so dangerous is that humans evolved to walk bipedally, for reasons of vision and locomotion. Bipedal stature makes the hips a lot more narrow, and in turn, gives the baby less room to escape the birth canal. Evolution decided that risk was worth it. The things that make humans have difficult births were the exact same things that allow us to help each other with them. So we traded safety for knowledge, and we assist each other because we're smart enough to do so. TL;DR- Your cat can have 4 kittens on her own just fine because the cost of that is cat-intelligence. The cost of human intelligence is a difficult birth and a long childhood. Reproduction and evolution think that this is a reasonable price.
[ "Human infants are also almost always born with assistance from other humans because of the way that the pelvis is shaped. Since the pelvis and opening of birth canal face backwards, humans have difficulty giving birth themselves because they cannot guide the baby out of the canal. Non-human primates seek seclusion when giving birth because they do not need any help due to the pelvis and opening being more forward. Human infants depend on their parents much more and for much longer than other primates. Humans spend a lot of their time caring for their children as they develop whereas other species stand on their own from when they are born. The faster an infant develops, the higher the reproductive output of a female can be. So in humans, the cost of slow development of their infants is that humans reproduce relatively slowly. This phenomenon is also known as cooperative breeding.\n", "Due to the somewhat unclear outlines of what, precisely, constitutes domestication, there are some species that may or may not be fully domesticated. There are also species that are extensively used or kept as pets by humans, but are not significantly altered from wild-type animals. Most animals on this second table are at least somewhat altered from wild animals by their extensive interactions with humans. Many could not be released into the wild, or are in some way dependent on humans.\n", "Evolutionary anthropologist Sarah Hrdy observes that only human mothers among great apes are willing to let another individual take hold of their own babies; further, we are routinely willing to let others babysit. She identifies lack of trust as the major factor preventing chimp, bonobo or gorilla mothers from doing the same: \"If ape mothers insist on carrying their babies everywhere ... it is because the available alternatives are not safe enough.\" The fundamental problem is that ape mothers (unlike monkey mothers who may often babysit) do not have female relatives nearby. The strong implication is that, in the course of \"Homo\" evolution, allocare could develop because \"Homo\" mothers did have female kin close by — in the first place, most reliably, their own mothers. Extending the Grandmother hypothesis, Hrdy argues that evolving \"Homo erectus\" females necessarily relied on female kin initially; this novel situation in ape evolution of mother, infant and mother's mother as allocarer provided the evolutionary ground for the emergence of intersubjectivity. She relates this onset of \"cooperative breeding in an ape\" to shifts in life history and slower child development, linked to the change in brain and body size from the 2 million year mark.\n", "Some animals reproduce with male and female; some animals change sex – they start female and they end male or vice-versa. Some fish do that. Some animals are hermaphrodites – they don't need anybody, they have both vaginas and penises. Then we have animals that don't need sex at all, they just clone themselves.\n", "The obstetrical dilemma is a hypothesis to explain why humans often require assistance from other humans during childbirth to avoid complications, whereas most non-human primates give birth alone with relatively little difficulty. The obstetrical dilemma posits that this is due to the biological trade-off imposed by two opposing evolutionary pressures in the development of the human pelvis. As human ancestor species (hominids) developed bipedal locomotion (the ability to walk upright), decreasing the size of the bony birth canal, they also developed ever larger skulls, which required a wider obstetrical pelvic area to accommodate this trend in hominid infants.\n", "However, there is disagreement over the definition of the term \"domestication\". One interpretation of \"domestication\" is that it must include physiological changes associated with being selectively bred in captivity, and not merely \"tamed.\" It has been noted that traditional peoples worldwide (both hunter-gatherers and horticulturists) routinely tame individuals from wild species, typically by hand-rearing infants whose parents have been killed, and these animals are not necessarily \"domesticated.\" \n", "The practice of keeping wild animals as pets threatens several species. Baby animals are typically captured and sold, which often requires killing the mother. Once in captivity and out of their natural habitat, many pets die or fail to reproduce. Affected populations include the Asiatic black bear, Malayan sun bear, white-handed lar, pileated gibbon and binturong.\n" ]
why do some acne medications cause an "initial breakout," making your skin worse, before making it better?
Hi, it’s because your cells are turning over rapidly and pushing acne that is below the surafce to the top. This is quite normal. Now if at any point your acne become severe and it seems that the breakout is not normal for YOU, go to a derm just to make sure you are not allergic to the product. It is a good idea to start slow and if the side effects are too much, cut back
[ "Acne usually flares up 2–3 weeks into the treatment and is usually mild and tolerable. Occasionally this flare-up is severe, necessitating oral antiobiotics such as erythromycin. A short course of oral prednisolone may be required. Some dermatologists favour a few weeks of pre-treatment with oral antibiotics before commencing isotretinoin to reduce the chance of a severe flare. A \"stepped\" course may also be used to reduce the chance of this initial flare, by which the initial dose is low (e.g. 0.5 mg/kg) and subsequently increased throughout the course.\n", "The US Food and Drug Administration (FDA) warned in 2014 that over-the-counter acne products containing benzoyl peroxide and/or salicylic acid, including Proactiv, can cause severe irritation, as well as rare but life-threatening allergic reactions. Consumers were advised to stop using the products if they experience hives or itching, and to seek emergency medical attention if they feel faint, or experience throat tightness, breathing problems, or swelling of the eyes, face, lips or tongue. The FDA noted that it remains unclear whether the reactions are caused by the active ingredients, inactive ingredients or a combination.\n", "For the treatment of acne, the most prescribed retinoid drug is 13-cis retinoic acid (isotretinoin). It reduces the size and secretion of the sebaceous glands. Although it is known that 40 mg of isotretinoin will break down to an equivalent of 10 mg of ATRA — the mechanism of action of the drug (original brand name Accutane) remains unknown and is a matter of some controversy. Isotretinoin reduces bacterial numbers in both the ducts and skin surface. This is thought to be a result of the reduction in sebum, a nutrient source for the bacteria. Isotretinoin reduces inflammation via inhibition of chemotactic responses of monocytes and neutrophils. Isotretinoin also has been shown to initiate remodeling of the sebaceous glands; triggering changes in gene expression that selectively induce apoptosis. Isotretinoin is a teratogen with a number of potential side-effects. Consequently, its use requires medical supervision.\n", "Several drugs may cause AC as a side effect, by various mechanisms, such as creating drug-induced xerostomia. Various examples include isotretinoin, indinavir, and sorafenib. Isotretinoin (Accutane), an analog of vitamin A, is a medication which dries the skin. Less commonly, angular cheilitis is associated with primary hypervitaminosis A, which can occur when large amounts of liver (including cod liver oil and other fish oils) are regularly consumed or as a result from an excess intake of vitamin A in the form of vitamin supplements. Recreational drug users may develop AC. Examples include cocaine, methamphetamines, heroin, and hallucinogens.\n", "In medicine, a drug eruption is an adverse drug reaction of the skin. Most drug-induced cutaneous reactions are mild and disappear when the offending drug is withdrawn. These are called \"simple\" drug eruptions. However, more serious drug eruptions may be associated with organ injury such as liver or kidney damage and are categorized as \"complex\". Drugs can also cause hair and nail changes, affect the mucous membranes, or cause itching without outward skin changes.\n", "It is important to recognize early that these drugs are causing DIL like symptoms and discontinue use of the drug. Symptoms of drug-induced lupus erythematosus generally disappear days to weeks after medication use is discontinued. Non-steroidal anti-inflammatory drugs (NSAIDs) will quicken the healing process. Corticosteroids may be used if more severe symptoms of DIL are present.\n", "Intramuscular progesterone often causes pain when injected. It irritates tissues and is associated with injection site reactions such as changes in skin color, pain, redness, transient indurations (due to inflammation), ecchymosis (bruising/discoloration), and others. Rarely, sterile abscesses can occur.\n" ]
why do professional swimmers wear 2 caps when competing?
So I was actually wondering this out loud the other day while I was watching the Olympics with my wife, and not 10 seconds later the commentator on the TV actually explained it. He said occasionally you will see some swimmers wear their goggles with the strap on the outside of their swimming cap, but most wear them with the strap on the inside. Some swimmers wear their goggles with the strap in contact with their hair, and then a single swim cap over top, but other swimmers may feel that their hair is too slippery for the goggles' strap to stay in place, so they wear a swim cap, then the goggles with the strap over top of the first swim cap, and then a second swim cap to make sure their goggles can't slip off. He seemed to know what he was talking about, so I trust his explanation.
[ "Unlike triathlons, which allow swimmers to wear wetsuits when the water is below a certain temperature (the standard is at the surface or up to for unofficial events.), most open water swim races either do not permit the use of wetsuits (usually defined as anything covering the body above the waist or below the knees), or put wetsuit-clad swimmers in a separate category and/or make them ineligible for race awards. This varies by locales and times of the year, where water temperatures are substantially below comfortable.\n", "Because open water swim areas are often cold and because wearing a wetsuit provides a competitive advantage, specialized triathlon wetsuits have been developed in a variety of styles to match the conditions of the water. For example, wetsuits that are sleeveless and cut above the knee are designed for warmer waters, while still providing buoyancy. Wetsuits are legal in sanctioned events at which the surface water temperature is or less. In non-sanctioned events or in \"age group\" classes where most racers are simply participating for the enjoyment of the sport instead of vying for official triathlon placing, wetsuits can often be used at other temperatures. Race directors will sometimes discourage or ban wetsuits if the water temperature is above due to overheating that can occur while wearing a wetsuit. Other rules have been implemented by race organizers regarding both wetsuit thickness as well as the use of \"swim skins;\" which need to be considered by those participating in future triathlons. Some triathlon sanctioning bodies have placed limits on the thickness of the wetsuit material. Under ITU and some national governing bodies' rules no wetsuit may have a thickness of more than .\n", "Swim briefs are worn by professional and recreational athletes in many water sports. They are the standard for competitive diving and water polo. They are preferred in competitive swimming for the reduction of the water's drag on the swimmer, although jammers and bodyskins are sometimes worn instead of the swim brief. Participants in sports that require a wetsuit such as waterskiing, scuba diving, surfing, and wakeboarding often wear swim briefs as an undergarment to the wetsuit.\n", "Some swimsuits are designed specifically for swimming competitions where they may be constructed of a special low resistance fabric that reduces skin drag. For some kinds of swimming and diving, special bodysuits called \"diveskins\" are worn. These suits are made from spandex and provide little thermal protection, but they do protect the skin from stings and abrasion. Most competitive swimmers also wear special swimsuits including partial bodysuits, racerback styles, jammers and racing briefs to assist their glide through the water thus gaining a speed advantage.\n", "A 'personal issue' wetsuit protects each diver from injury and retains enough warmth to withstand even the coldest water. An inflatable 'Fenzy' assists ascent to the surface through a compressed air cylinder, which automatically inflates the vest. Standard diving accessories include swim fins, facemask, diving knives, weight belt and a lifeline between the diver and the on-land attendant.\n", "Some triathlon wetsuits go further, and use rubber-molding and texturing methods to roughen up the surface of the suit on the forearms, to increase forward drag and help pull the swimmer forwards through the water. Extremely thin 1 mm neoprene is also often used in the under-arm area, to decrease stretch resistance and reduce strain on the swimmer when they extend their arms out over their head.\n", "Unlike regular swimsuits, which are designed mainly for the aesthetic appearances, swimsuits designed to be worn during competitions are manufactured to assist the athlete in swim competitions. They reduce friction and drag in the water, increasing the efficiency of the swimmer's forward motion. The tight fits allow for easy movement and are said to reduce muscle vibration, thus reducing drag. This also reduces the possibility that a high forwards dive will remove a divers swimwear. Starting around 2000, in an effort to improve the effectiveness of the swimsuits, engineers have taken to designing them to replicate the skin of sea-based animals, sharks in particular.\n" ]
Were number systems written down before words and language?
Sumerian cuneiform has its origin in accounting systems, particularly those used by temples to keep track of goods like cattle and grain. A sign could represent a type of good, and the number with a system of tick marks next to it. Cuneiform symbols originated as mnemonics/pictograms, which were subsequently abstracted, and the representational/mnemonic function of the symbols became more sophisticated over time, developing into a more complete system. [This website](_URL_0_) has a more detailed explanation of how the system developed. Some kind of abstract numerical representation system does seem to predate what we think of as written language, but the development of written language was very gradual--it wouldn't be accurate to mark a strict cutoff before which proto-writing could be considered merely a mnemonic system and after which it could be considered "fully-fledged writing". John Hayes, in the *Manual of Sumerian Grammar and Texts*, points out that written Sumerian in all likelihood is an incomplete representation of the language as it was spoken, especially phonologically. Writing may have developed quite differently when it arose in other places--the earliest Chinese writing is from inscriptions on bones used in pyromancy, not accounting. The exact relationship of Egyptian hieroglyphics to cuneiform is still debated, and I don't know if anything is known about the development of Mayan writing, so it's hard to say if the method of development of writing in Mesopotamia out of an accounting system represents a general tendency or a peculiarity. And, of course, the written accounting tools used by the ancient Mesopotamians were still pretty basic--mathematical notation and double-entry bookkeeping are both much more recent inventions. What they had were pretty much just counting systems. Hope some of that helps.
[ "Traditionally, certain letters were rendered differently according to a set of rules. In particular, those letters that began sentences or nouns were made larger and often written in a distinct script. There was no fixed capitalisation system until the early 18th century. The English language eventually dropped the rule for nouns, while the German language keeps it.\n", "The system was a spoken one in the sense that consonants and vowels which are not vocalised have no numerical value. The system is a place-value system with zero. In fact many different \"words\" could represent the same number and this was highly useful for works written in verse.\n", "Phonemes are considered to be the basis for alphabetic writing systems. In such systems the written symbols (graphemes) represent, in principle, the phonemes of the language being written. This is most obviously the case when the alphabet was invented with a particular language in mind; for example, the Latin alphabet was devised for Classical Latin, and therefore the Latin of that period enjoyed a near one-to-one correspondence between phonemes and graphemes in most cases, though the devisers of the alphabet chose not to represent the phonemic effect of vowel length. However, because changes in the spoken language are often not accompanied by changes in the established orthography (as well as other reasons, including dialect differences, the effects of morphophonology on orthography, and the use of foreign spellings for some loanwords), the correspondence between spelling and pronunciation in a given language may be highly distorted; this is the case with English, for example.\n", "In ancient times, numbers in Latin were written only with letters. Today, the numbers can be written with the Arabic numbers as well as with Roman numerals. The numbers 1, 2 and 3 and every whole hundred from 200 to 900 are declined as nouns and adjectives, with some differences.\n", "The principle behind alphabetic writing systems is that the letters (graphemes) represent phonemes. However, in many orthographies based on such systems the correspondences between graphemes and phonemes are not exact, and it is sometimes the case that certain spellings better represent a word's morphophonological structure rather than the purely phonological. An example of this is that the English plural morpheme is written \"-s\" regardless of whether it is pronounced as or ; we write \"cats and \"dogs, not \"dogz\".\n", "This alphabetic system operates on the additive principle in which the numeric values of the letters are added together to obtain the total. For example, 241 was represented as  (200 + 40 + 1). (It was not always the case that the numbers ran from highest to lowest: a 4th-century BC inscription at Athens placed the units to the left of the tens. This practice continued in Asia Minor well into the Roman period.) In ancient and medieval manuscripts, these numerals were eventually distinguished from letters using overbars: , , , etc. In medieval manuscripts of the Book of Revelation, the number of the Beast 666 is written as  (600 + 60 + 6). (Numbers larger than 1,000 reused the same letters but included various marks to note the change.) Fractions were indicated as the denominator followed by a \"keraia\" (ʹ); γʹ indicated one third, δʹ one fourth and so on. As an exception, special symbol ∠ʹ indicated one half. These fractions were additive (also known as Egyptian fractions); for example indicated .\n", "In some contexts, numbers and letters are used not so much as a basis for establishing an ordering, but as a means of labeling items that are already ordered. For example, pages, sections, chapters, and the like, as well as the items of lists, are frequently \"numbered\" in this way. Labeling series that may be used include ordinary Arabic numerals (1, 2, 3, ...), Roman numerals (I, II, III, ... or i, ii, iii, ...), or letters (A, B, C, ... or a, b, c, ...). (An alternative method for indicating list items, without numbering them, is to use a bulleted list.)\n" ]
what defines a religion from mythology?
A mythology is a set of traditional stories. The Jewish mythology is the Old Testament, the Christian mythology is the Old and New Testaments, the Greek mythology is... well, Greek mythology. Mythology doesn't have anything to do with whether the stories are true or not, it's an agnostic term. It simply says "these are stories that people tell each other." A religion is a set of beliefs that is accepted on faith by its followers. Religions often incorporate mythologies as part of their belief set, though it's not totally necessary. The point here is that "religion" and "mythology" are not interchangeable terms. One is a set of stories, one is a set of beliefs. It's confusing because the colloquial usage for "mythology" (and "myth" especially) often *implies* that it's something that's untrue, so many religious people avoid calling the stories from their faith "myths." They are, though, by the strictest technical definition.
[ "Mythology is the main component of Religion. It refers to systems of concepts that are of high importance to a certain community, making statements concerning the supernatural or sacred. Religion is the broader term, besides mythological system, it includes ritual. A given mythology is almost always associated with a certain religion such as Greek mythology with Ancient Greek religion. Disconnected from its religious system, a myth may lose its immediate relevance to the community and evolve—away from sacred importance—into a legend or folktale. \n", "All over the world, religion is formed around a divine, supernatural figure. While the idea of the divine, supernatural figure varies from religion to religion, each one is framed around different concepts of what it means to be male and female. Furthermore, the religion of a culture usually directly corresponds or is influenced by the culture's gender structure, like the family structures and/or the state. Therefore the religious structure and the gender structure work together to form and define a culture, creating the defining structures of equality and uniformity.\n", "A religion is a system of human thought which usually includes a set of narratives, symbols, beliefs and practices that give meaning to the practitioner's experiences of life through reference to a higher power, God or gods, or ultimate truth.\n", "Christian mythology is the body of myths associated with Christianity. The term encompasses a broad variety of legends and stories, especially those considered sacred narratives. Mythological themes and elements occur throughout Christian literature, including recurring myths such as ascending to a mountain, the \"axis mundi\", myths of combat, descent into the Underworld, accounts of a dying-and-rising god, flood stories, stories about the founding of a tribe or city, and myths about great heroes (or saints) of the past, paradises, and self-sacrifice.\n", "\"Religion\" is a belief concerning the supernatural, sacred, or divine, and the moral codes, practices, values, and institutions associated with such belief, although some scholars, such as Durkheim, would argue that the supernatural and the divine are not aspects of all religions. Religious beliefs and practices may include the following: a deity or higher being, eschatology, practices of worship, practices of ethics and politics. Some religions do not include all these features. For instance, belief in a deity is not essential to Buddhism.\n", "A \"religion\" is a set of beliefs and practices, often centered upon specific supernatural and/or moral claims about reality, the cosmos, and human nature, and often codified as prayer, ritual, and law. Religion also encompasses ancestral or cultural traditions, writings, history, and mythology, as well as personal faith and mystic experience. The term \"religion\" refers to both the personal practices related to communal faith and to group rituals and communication stemming from shared conviction.\n", "Eliade argues that religion is based on a sharp distinction between the sacred (God, gods, mythical ancestors, etc.) and the profane. According to Eliade, for traditional man, myths describe \"breakthroughs of the sacred (or the 'supernatural') into the World\"—that is, hierophanies. \n" ]
how did they predict existence of subatomic particles, black holes, multiple universes using maths and equations?
> subatomic particles They didn't predict those with math. They were largely experimentally discovered. Thomson's cathode ray experiment found that there were negatively charged subatomic particles we now call electrons, for example. > black holes These were first thought of from Newtonian mechanics by Laplace (and others?), but it was really just an idea. After Einstein produced the general relativity theory, Schwarzschild predicted the limit at which a black hole could form. > multiple universes Those aren't mathematically predicted either, they're really outside the realm of science.
[ "Improved understanding of the world of particles prompted physicists to make bold predictions, such as Dirac's positron in 1928 (founded on the Dirac Sea model) and Pauli's neutrino in 1930 (founded on conservation of energy and angular momentum in beta decay). Both were later confirmed.\n", "The Standard Model of particle physics is a theory which describes three of the four known fundamental interactions between the elementary particles that make up all matter. This theory allows predictions to be made about how particles will interact under many conditions. The energy per particle in a supernova is typically 1–150 picojoules (tens to hundreds of MeV). The per-particle energy involved in a supernova is small enough that the predictions gained from the Standard Model of particle physics are likely to be basically correct. But the high densities may require corrections to the Standard Model. In particular, Earth-based particle accelerators can produce particle interactions which are of much higher energy than are found in supernovae, but these experiments involve individual particles interacting with individual particles, and it is likely that the high densities within the supernova will produce novel effects. The interactions between neutrinos and the other particles in the supernova take place with the weak nuclear force, which is believed to be well understood. However, the interactions between the protons and neutrons involve the strong nuclear force, which is much less well understood.\n", "Several of Loeb's early predictions were confirmed in recent years. In 1992, he suggested with Andy Gould that exoplanets could be detected through gravitational microlensing, a technique that is routinely used these days. In 1993, he proposed the use of the C+ fine-structure line to discover galaxies at high redshifts, as done routinely now. In 2005, he predicted in a series of papers with his postdoc at the time, Avery Broderick, how a hot spot in orbit around a black hole would appear; their predictions were confirmed in 2018 by the GRAVITY instrument on the VLT which observed a circular motion of the centroid of light of the black hole at the center of the Milky Way, SgrA*. In 2009, Broderick and Loeb predicted the shadow of the black hole in the giant elliptical galaxy M87, which was imaged in 2019 by the Event Horizon Telescope. In 2013, a report was published on the discovery of the \"Einstein Planet\" Kepler 76b, the first Jupiter size exoplanet identified through the detection of relativistic beaming of its parent star, based on a technique proposed by Loeb and Gaudi in 2003. In addition, a pulsar was discovered around the supermassive black hole, SgrA*, following a prediction by Pfahl and Loeb in 2004.\n", "While it is possible to assume that quantum mechanical predictions are due to nonlocal, hidden variables, and in fact David Bohm invented such a formulation, this resolution is not satisfactory to the vast majority of physicists. The question of whether a random outcome is predetermined by a nonlocal theory can be philosophical, and it can be potentially intractable. If the hidden variables are not constrained, they could just be a list of random digits that are used to produce the measurement outcomes. To make it sensible, the assumption of nonlocal hidden variables is sometimes augmented by a second assumption—that the size of the observable universe puts a limit on the computations that these variables can do. A nonlocal theory of this sort predicts that a quantum computer would encounter fundamental obstacles when attempting to factor numbers of approximately 10,000 digits or more; a potentially achievable task in quantum mechanics.\n", "Other theories of the very early universe have been proposed that are claimed to make similar predictions, such as the brane gas cosmology, cyclic model, pre-big bang model and holographic universe, but they remain nascent and are not widely accepted. Some theories, such as cosmic strings, have largely been refuted by increasingly precise data.\n", "In experimental physics, there are always observational errors determining variables such as positions and velocities. So perfect prediction is \"practically\" impossible. Moreover, in modern quantum mechanics, Werner Heisenberg's indeterminacy principle puts limits on the accuracy with which such quantities can be known. So such perfect predictability is also \"theoretically\" impossible. \n", "From the time of Newton until about 1890, it was generally believed that if one knows the initial state of a system with great accuracy, and if all the forces acting on the system can be formulated with equal accuracy, it would be possible, in principle, to make predictions of the state of the universe for an infinitely long time. The limits to such predictions in physical systems became clear as early as 1893 when Henri Poincaré showed that in the three-body problem in astronomy, small changes to the initial state could result in large changes in trajectories during the numerical integration of the equations.\n" ]
if your car gets stolen and your insurance covers it, what happens if the stolen car gets found after you have already gotten a new one?
Usually, it belongs to your insurance company now. They'll take full ownership of the stolen vehicle. This should be stated in your insurance policy.
[ "Some drivers believe that a new car is in greater danger than a used car of getting into an accident or having a collision. Some drivers will leave change under their seats. Others use one coin to scratch the car, based on the (false) belief that since the car is new and nothing has happened to it yet, the chances of something bad happening to the car is greater when compared to a used car which already has its fair share of dents and scratches. In hopes of preventing a high damaging accident, they will place a small nick or scratch on the car in an area where it will not be seen. The inside of the wheel well is one commonly scratched area.\n", "Often car rental and car leasing companies will require a deposit to protect themselves against possible damage to the car. Once the car is returned, it is checked for any possible damage, and if damage is found, funds are deducted from the deposit to cover the repairs and the loss of value.\n", "BULLET::::1. A deterrent to auto theft: If a vehicle is a complete loss due to an accident, its serial number (VIN, Vehicle identification number) and registration documents could still be of potential value to persons dealing in stolen cars. The diminished sale value of a title branded vehicle reduces the profitability of switching the registration and VIN from an accident vehicle to that of a stolen vehicle of the same make and model/year, in an attempt to register it as a rebuilt car and sell it.\n", "In many jurisdictions a vehicle designated as a total loss is sold by insurance companies to general public, auto dealers, auto brokers, or auto wreckers. The metrics insurance companies use to make the decision include the cost of the repairs needed plus the value of the remaining parts, added to the cost of reimbursing the driver for a rental while the car in question is repaired. If this figure exceeds the value of the car after it is repaired, the vehicle is deemed a total loss. In most jurisdictions, a decision by an insurer to write off a vehicle results in vehicle title branding, marking the car as \"salvage\" or (if repaired and reinspected under subsequent ownership) \"rebuilt\". If the vehicle is not severely damaged, however, it can be restored to its original condition. After a government approved inspection, the vehicle can be put back on the road. The inspection process may not attempt to assess the quality of the repairs. This function will be relegated to a professional mechanic or inspector. However, if the vehicle is severely damaged as per standards set by state or provincial governments, the vehicle is dismantled by an auto wrecker and is sold as parts or scrapped.\n", "In \"Commercial Union v Lotter\", the buyer of a luxury motor vehicle did not disclose to the insurer that the vehicle had been stolen from another country. When the vehicle was stolen again, the insurance company repudiated the claim. The court upheld the company's repudiation on the basis that material facts had not been disclosed. The insurance company argued that its right of subrogation was diminished by the fact that the vehicle in question was a stolen vehicle when the insurance policy was taken out: The insured had no title to the vehicle, so the insurance company could not sue a negligent third party, in terms of its right of subrogation, for the full costs of repairing any damage to the vehicle.\n", "BULLET::::- Salvage: vehicles that have been in accidents, floods, fires or recovered thefts that have been purchased by insurance companies. The insurance companies sell these vehicles to dealers or body shops who will fix them and resell them, or auto recyclers who will part out the remaining parts of the vehicle that haven't been damaged.\n", "Upon paying the claim, the insurer may offer to return the vehicle to the owner as an insurance buy-back, in which case the owner is responsible for having the repairs made and having the car inspected by a State-designated facility. Depending on the state, this inspection may remove the salvage brand from the vehicle's title. The exact percentage of value that triggers the decision to total the vehicle is guided by applicable laws and regulations. The damage estimate is calculated at retail repair rates, which may be more than the cost of wholesale repair. Vehicles that are not bought back are auctioned as salvage to an auto recycler or a rebuilder and given a salvage title.\n" ]
what is the difference between normal steel and galvanized steel?
Galvanizing is a process of adding a zinc coat to the steel. It should make the nuts last longer and prevents rust. The actual process is slightly more complicated but this is essentially it.
[ "Electrical steel is one material that uses decarburization in its production. To prevent the atmospheric gases from reacting with the metal itself, electrical steel is annealed in an atmosphere of nitrogen, hydrogen, and water vapor, where oxidation of the iron is specifically prevented by the proportions of hydrogen and water vapor so that the only reacting substance is carbon being made into carbon monoxide.\n", "Steel is an alloy of iron and carbon, and sometimes other elements. Because of its high tensile strength and low cost, it is a major component used in buildings, infrastructure, tools, ships, automobiles, machines, appliances, and weapons.\n", "Steel - basically a mixture of iron and carbon, although other metals may be added to change the characteristic of the steel (add chrome, get stainless steel; add nickel, get armor plate). The carbon content of steel is closely controlled in its manufacture - the more carbon the \"stronger\" the steel. Steel varies from iron in that it can be hardened. Heat steel red-hot and quench it and it gets hard, heat iron red hot and quench it and it gets cold.\n", "Electrical steel (lamination steel, silicon electrical steel, silicon steel, relay steel, transformer steel) is an iron alloy tailored to produce specific magnetic properties: small hysteresis area resulting in low power loss per cycle, low core loss, and high permeability.\n", "BULLET::::- Steel is a metal alloy whose major component is iron, and is the usual choice for metal structural building materials. It is strong, flexible, and if refined well and/or treated lasts a long time.\n", "While ferrocerium-and-steels function in a similar way to natural flint-and-steel in fire starting, ferrocerium takes on the role that steel played in traditional methods: when small shavings of it are removed quickly enough the heat generated by friction is enough to ignite those shavings, converting the metal to the oxide, i.e., the sparks are tiny pieces of burning metal. The sparking is due to cerium's low ignition temperature of between . About 700 tons were produced in 2000.\n", "Steel is a ductile material, which will behave elastically until it reaches yield (point 2 on the stress–strain curve), when it becomes plastic and will fail in a ductile manner (large strains, or extensions, before fracture at point 3 on the curve). Steel is equally strong in tension and compression.\n" ]
how are the "dark triad" traits in psychology different from each other?
Psychopathy is a lack of empathy & understanding of how people feel. Narcissism is placing your own needs & wants over others. Machiavellianism is lying & manipulating people to achieve your goals. They're called the "dark triad" because they do so often overlap & feed into each other. It's easy to be narcissistic if you don't care about how people feel. It's easy to manipulate people if you feel your needs are more important than theirs.
[ "There is a good deal of conceptual and empirical overlap between the dark triad traits. For example, researchers have noted that all three traits share characteristics such as a lack of empathy, interpersonal hostility, and interpersonal offensiveness. Likely due in part to this overlap, a number of measures have recently been developed that attempt to measure all three dark triad traits simultaneously, such as the Dirty Dozen and the short dark triad (SD3).\n", "The dark triad is made up of three personality traits, psychopathy, narcissism and Machiavellianism and was proposed by Paulhus and Williams (2002) . The three traits are exploitative in nature and are used for sexually coercive behaviours, useful in the seduction process. Typically these three traits are deemed maladaptive for the individual and society. Nevertheless, these traits have been found to be adaptive in an exploitative strategy in short term mating. Dark triad traits are adaptive for an unrestricted sociosexuality and promiscuous behaviours. The three traits are associated with impulsivity, manipulative behaviours and lack of empathy. These personality traits would be useful in seducing a partner for a short term encounter. From an evolutionary perspective, these would have been particularly beneficial to our ancestral males who wanted to increase their reproductive success, through seducing many women and therefore increasing their chance of passing on their genes. These particular traits may be used as a tactic for increasing success in mating.\n", "Paulhus and Williams (2002) coined the term \"dark triad\" in referring to three socially aversive personalities: machiavellianism, narcissism, and psychopathy. The research showed both similarities and differences among the three constructs. Their distinctiveness was confirmed in studies of associations with impulsivity, aggression, body modification, mate choice, sexual deviancy, scholastic cheating, revenge, and the personality of stalkers. A fourth member, everyday sadism, was recently added to the pantheon of dark personalities. Questionnaire measures are available in a chapter by Paulhus and Jones (2015).\n", "Psychopathy, narcissism and Machiavellianism, three personality traits that are together referred to as the dark triad, share certain characteristics, such as a callous-manipulative interpersonal style. The dark tetrad refers to these traits with the addition of sadism.\n", "Research on the dark triad is used in applied psychology, especially within the fields of law enforcement, clinical psychology, and business management. People scoring high on these traits are more likely to commit crimes, cause social distress and create severe problems for an organization, especially if they are in leadership positions (for more information, see psychopathy, narcissism, and Machiavellianism in the workplace). They also tend to be less compassionate, agreeable, empathetic, satisfied with their lives, and likely to believe they and others are good.\n", "The dark triad of personality consists of psychopathy, Machiavellianism and narcissism. Psychopathy is identified by characteristics such as remorselessness, antisociality and selfishness. Machiavellianism consists of selfishness, in that one will focus on their own needs, primarily by manipulating others. Narcissism can also be defined as selfishness, but is different as this person would consider themselves of a higher importance than those around them. However, these constructs are said to be not fully represented in common five-factor models of personality. The Dark Triad can be conceptualized as being on the opposite pole of Honesty-Humility (Sincere, Faithful, Loyal etc.), which would mean that low levels of Honesty-Humility corresponds to higher levels of psychopathy, Machiavellianism and/or narcissism. The Dark Triad personality constructs tend to only correlate with disagreeableness on the Big Five Inventory, otherwise they are represented inconsistently on measures of the Big Five traits. For that reason, several researchers have used the HEXACO model to gain a more detailed understanding of the personality characteristics of individuals who exhibit traits/behaviours that would be considered along the Dark Triad dimension.\n", "found to have substantial genetic components. It has also been found that the observed relationships among the dark triad, and among the dark triad and the Big Five, are strongly driven by individual differences in genes. However, while psychopathy (h = 0.64) and narcissism (h = 0.59) both have a relatively large heritable component, Machiavellianism (h = 0.31) while also moderately influenced by genetics, has been found to be less heritable than the other two traits.\n" ]
Why don't we launch spacecraft using magnets?
To get into low earth orbit with a rail gun would require an exit velocity so high that whatever you are launching would burn up immediately. It has been suggested as an efficient way to launch things from bodies without an atmosphere.
[ "The main disadvantage of magnetorquers is that very high magnetic flux densities are needed if large craft have to be turned very fast. This either necessitates a very high current in the coils, or much higher ambient flux densities than are available in Earth orbit. Consequently, the torques provided are very limited and only serve to accelerate or decelerate the change in a spacecraft's attitude by minute amounts. Over time active control can produce very fast spinning even here, but for accurate attitude control and stabilization the torques provided often aren't enough. To overcome this, magnetorquer are often combined with reaction wheels.\n", "All current spacecraft use chemical rockets (bipropellant or solid-fuel) for launch, though some (such as the Pegasus rocket and SpaceShipOne) have used air-breathing engines on their first stage. Most satellites have simple reliable chemical thrusters (often monopropellant rockets) or resistojet rockets for orbital station-keeping and some use momentum wheels for attitude control. Soviet bloc satellites have used electric propulsion for decades, and newer Western geo-orbiting spacecraft are starting to use them for north-south stationkeeping and orbit raising. Interplanetary vehicles mostly use chemical rockets as well, although a few have used ion thrusters and Hall effect thrusters (two different types of electric propulsion) to great success.\n", "In theory, it is possible for a magnetic sail to launch directly from the surface of a planet near one of its magnetic poles, repelling itself from the planet's magnetic field. However, this requires the magnetic sail to be maintained in its \"unstable\" orientation. A launch from Earth requires superconductors with 80 times the current density of the best known high-temperature superconductors.\n", "Due to their reliability, magnetos are used as ignition systems on aircraft. They are also used on machinery that do not have a separate electric supply or battery. They are also used on drag race cars because they offer a weight advantage over systems that utilize a distributor and battery.\n", "A magnetic sail or magsail is a proposed method of spacecraft propulsion which would use a static magnetic field to deflect charged particles radiated by the Sun as a plasma wind, and thus impart momentum to accelerate the spacecraft. A magnetic sail could also thrust directly against planetary and solar magnetospheres.\n", "Due to weight issues such as shielding, many nuclear propulsion systems are unable to lift their own weight, and hence are unsuitable for launching to orbit. However some designs such as the Orion project and some nuclear thermal designs do have a thrust to weight ratio in excess of 1, enabling them to lift off. Clearly one of the main issues with nuclear propulsion would be safety, both during a launch for the passengers, but also in case of a failure during launch. No current program is attempting nuclear propulsion from Earth's surface.\n", "On rockets with a solid motor upper stage, spin stabilization is used to keep the motor from drifting off course as they don't have their own thrusters. Usually small rockets are used to spin up the spacecraft and rocket then fire the rocket and send the craft off.\n" ]
how does winrar/7-zip just make my files smaller?
Just about any set of information contains some repeating information. My paragraph contains the word contains several times. One means to reduce the size would be to put a short unique identifier in place of each word "contains" and a reference that the identifier means contains. So a very simple compression scheme for the above paragraph might look like: Just about any set of ! @ some repeating !. My paragraph @ the word @ several times. One means to reduce the size would be to put a short unique # in place of each word "@" and a reference that the # means @. !=information @=contains \#=identifier That's a quick 15% reduction in the characters (counting the replacement information we added). If we did the same with parts of words, or shorter binary strings, we could further reduce the space. It's easy to automatically replace the characters with the original words when you want the original file back. Compression schemes do something similar with the binary data that makes up all computer files. That's why compression works better on data with lots of repeating information (like text or a bitmap image) rather than data with little repeating information (like a jpg).
[ "7-Zip comes with a file manager along with the standard archiver tools. The file manager has a toolbar with options to create an archive, extract an archive, test an archive to detect errors, copy, move, and delete files, and open a file properties menu exclusive to 7-Zip. The file manager, by default, displays hidden files because it does not follow Windows Explorer's policies. The tabs show name, modification time, original and compressed sizes, attributes, and comments (4DOS codice_3 format).\n", "7-Zip is a free and open-source file archiver, a utility used to place groups of files within compressed containers known as \"archives\". It is developed by Igor Pavlov and was first released in 1999. 7-Zip uses its own 7z archive format, but can read and write several other archive formats. The program can be used from a command-line interface as the command p7zip, or through a graphical user interface that also features shell integration. Most of the 7-Zip source code is under the GNU LGPL license; the unRAR code, however, is under the GNU LGPL with an \"unRAR restriction\", which states that developers are not permitted to use the code to reverse-engineer the RAR compression algorithm.\n", "Although the original 7-Zip program, which implements LZMA2 compression, is able to produce small files at the cost of speed, it also created its own unique archive format which was made primarily for Windows and did not support Unix functionality.\n", "On the other hand, getting a single file out of a solid archive originally required processing all the files before it, so modifying solid archives could be slow and inconvenient. Later versions of 7-zip use a variable solid block size, so that only a limited amount of data must be processed in order to extract one file. Parameters control the maximum solid block window size, the number of files in a block, and whether blocks are separated by file extension.\n", "By default, 7-Zip creates 7z-format archives with a codice_1 file extension. Each archive can contain multiple directories and files. As a \"container\" format, security or size reduction are achieved using a stacked combination of filters. These can consist of pre-processors, compression algorithms, and encryption filters.\n", "ZIP is an archive file format that supports lossless data compression. A ZIP file may contain one or more files or directories that may have been compressed. The ZIP file format permits a number of compression algorithms, though DEFLATE is the most common. This format was originally created in 1989 and released to the public domain on February 14, 1989 by Phil Katz, and was first implemented in PKWARE, Inc.'s PKZIP utility, as a replacement for the previous ARC compression format by Thom Henderson. The ZIP format was then quickly supported by many software utilities other than PKZIP. Microsoft has included built-in ZIP support (under the name \"compressed folders\") in versions of Microsoft Windows since 1998. Apple has included built-in ZIP support in Mac OS X 10.3 (via BOMArchiveHelper, now Archive Utility) and later. Most have built in support for ZIP in similar manners to Windows and Mac OS X.\n", "When writing a new file of a known size, if there are any empty holes that are larger than that file, the operating system can avoid data fragmentation by putting the file into any one of those holes. There are a variety of algorithms for selecting which of those potential holes to put the file; each of them is a heuristic approximate solution to the bin packing problem. The \"best fit\" algorithm chooses the smallest hole that is big enough. The \"worst fit\" algorithm chooses the largest hole. The \"first-fit algorithm\" chooses the first hole that is big enough. The \"next fit\" algorithm keeps track of where each file was written.\n" ]
model-view-controller pattern
MVC isn't that hard. You have the view (or views). This is everything you see. It's a lot of rules about how to show the info. Then you have the model, which is the data. Then you have the controller. The controller takes the data and moves it into the view. If something changes, you tell the controller and he updates the correct places. Example: View: TextBox age: Position left Slider ageSlider: position right %%could have separated them into two views if I wanted to onSliderUpdate( NotifyControler(newAge) ) Model: Person{ int age. } Controller: OnNotify(int age) model.age = age for each view updateView
[ "Model–View–Controller (usually known as MVC) is an architectural pattern commonly used for developing user interfaces that divides an application into three interconnected parts. This is done to separate internal representations of information from the ways information is presented to and accepted from the user. The MVC design pattern decouples these major components allowing for code reuse and parallel development.\n", "BULLET::::- Model-View-Controller is an architectural design pattern which divides an application into separate parts which communicate with each other in a specific way. The goal is to separate data model (model), graphical user interface (view) and business logic (controller). GEF uses the MVC pattern extensively.\n", "Traditional model-view-controller (MVC) frameworks work in terms of whole requests and whole pages. In each request cycle, the incoming request is mapped to a method on a \"controller\" object, which then generates the outgoing response in its entirety, usually by pulling data out of a \"model\" to populate a \"view\" written in specialized template markup. This keeps the application's flow-of-control simple and clear, but can make code reuse in the controller difficult.\n", "A \"model\" represents the state of a particular aspect of the application. A \"controller\" handles interactions and updates the model to reflect a change in state of the application, and then passes information to the view. A \"view\" accepts necessary information from the controller and renders a user interface to display that information.\n", "In the field of control engineering, a map-based controller is a controller whose outputs are based on values derived from a pre-defined lookup table. The inputs to the controller are usually values taken from one or more sensors and are used to index the output values in the lookup table. By effectively placing the transfer function as discrete entries within a lookup table, engineers free to modify smaller sections or update the whole list of entries as required.\n", "Hierarchical model–view–controller (HMVC) is a software architectural pattern, a variation of model–view–controller (MVC) similar to presentation–abstraction–control (PAC), that was published in 2000 in an article in JavaWorld Magazine, the authors apparently unaware of PAC, which was published 13 years earlier.\n", "The controller has some oversight in that it selects first the model and then the view, realizing an approval mechanism by the controller. The model prevents the view from accessing the data source directly.\n" ]
why do people like to believe in god
I imagine you'll get a lot of responses along the lines of "they need easy comfort" or "they were brainwashed as children". I don't think these are untrue. But they're too simplified, even for ELI5. As far as comfort goes, it's not just comfort people are looking for. It's order. Humans love patterns and we really love authority. We're still very tribal in nature and that concept of a central authority for us isn't just appealing, it's a desire that is a base part of out nature. This is why you often see religions in which gods or a god promise a specific group of people that they are special. This god was part of a tribe. The leader of that tribe. Gods are a powerful explanation and a powerful authority. As history shows, our commitment to that authority and the sense of community is so strong that many of us simply can't let go of. This is the part where the brainwashing children comes in. I wouldn't say it's intentional, at least not in any kind of malicious way (fundamentalists and extremists notwithstanding, they're a totally different story). It's more that we develop our connection to tribe/community very very strongly as children. And if your tribe follows a god as a leader, you will as well.
[ "My faith is based on a constant search but I don't search frantically. It's no use to rush out into space to search for some gods there, if they want to have anything to do with me, they will come. I have often become aware of them, but I don't rush after them or shout at them. I have gotten to know them a bit in myself and also in other people. ... Primarily it is the effects of the great force felt by everyone that make me religious. ... The most remarkable thing about faith is that it gives us growth, the possibility to grow and thrive. And humility cannot be neglected. Without it we cannot live to any useful degree, though of course it has its particular place. But a man who is completely without it is a madman.\n", "We believe in the one High God of love who created the beautiful world and everything good in it. He created man and wanted man to be happy in the world. God loves the world and every nation and tribe in the world. We have known this God in darkness, and we now know God in the light. God promised in his book the Bible that he would save the world and all the nations and tribes.\n", "I cannot understand why we idle discussing religion. If we are honest—and scientists have to be—we must admit that religion is a jumble of false assertions, with no basis in reality. The very idea of God is a product of the human imagination. It is quite understandable why primitive people, who were so much more exposed to the overpowering forces of nature than we are today, should have personified these forces in fear and trembling. But nowadays, when we understand so many natural processes, we have no need for such solutions. I can't for the life of me see how the postulate of an Almighty God helps us in any way. What I do see is that this assumption leads to such unproductive questions as why God allows so much misery and injustice, the exploitation of the poor by the rich and all the other horrors He might have prevented. If religion is still being taught, it is by no means because its ideas still convince us, but simply because some of us want to keep the lower classes quiet. Quiet people are much easier to govern than clamorous and dissatisfied ones. They are also much easier to exploit. Religion is a kind of opium that allows a nation to lull itself into wishful dreams and so forget the injustices that are being perpetrated against the people. Hence the close alliance between those two great political forces, the State and the Church. Both need the illusion that a kindly God rewards—in heaven if not on earth—all those who have not risen up against injustice, who have done their duty quietly and uncomplainingly. That is precisely why the honest assertion that God is a mere product of the human imagination is branded as the worst of all mortal sins.\n", "In his book \"Why Would Anyone Believe in God?\" he suggests that \"belief in God is an almost inevitable consequence of the kind of minds we have. Most of what we believe comes from mental tools working below our conscious awareness. And what we believe consciously is in large part driven by these unconscious beliefs.\" and \"that beliefs in gods match up well with these automatic assumptions; beliefs in an all-knowing, all-powerful God match up even better.\"\n", "Justin L. Barrett in \"Why Would Anyone Believe in God?\" (2004) suggests that belief in God is natural because it depends on mental tools possessed by all human beings. He suggests that the structure and development of human minds make belief in the existence of a supreme god (with properties such as being superknowing, superpowerful and immortal) highly attractive. He also compares belief in God to belief in other minds, and devotes a chapter to looking at the evolutionary psychology of atheism. He suggests that one of the fundamental mental modules in the brain is the Hyperactive Agency Detection Device (HADD), another potential system for identifying danger. This HADD may confer a survival benefit even if it is over-sensitive: it is better to avoid an imaginary predator than be killed by a real one. This would tend to encourage belief in ghosts and in spirits.\n", "\"We do not always believe that God can be so concrete in our daily lives, so close and real, and even less so that He makes himself present by acting through someone we know, such as a neighbor, a friend, a relative. It is not uncommon for us to behave like the neighbors of Nazareth, preferring a God at a distance: magnificent, good, generous but distant and not bothering. Because a close God in everyday life, friend and brother asks us to learn closeness, daily presence and, above all, fraternity. (...) God is real, because love is real; God is concrete because love is concrete.\"\n", "We also have a religion, which was given to our forefathers, and has been handed down to us their children. We worship in that way. It teaches us to be thankful for all the favors we receive; to love each other, and to be united. We never quarrel about religion.\n" ]
why do people go to the bathroom on a pretty regular schedule?
Because generally your food/drink intake is fairly regular as well. You have coffee every day at 8am. Lunch at noon, dinner at 7 etc.
[ "BULLET::::- Too many people using a toilet: This is especially true in case of shared or public toilets. If too many people want to use a toilet at the same time, then some people may go outside to defecate instead of waiting. In some cases, people might not be able to wait due to diarrhea (or result of an Irritable Bowel Syndrome emergency).\n", "Other rooms are meant to promote comfort and cleanliness, such as the toilet and bathroom, which may be combined or which may be in separate rooms. The public equivalent is the restroom, which usually features a toilet and handwashing facilities, but not usually a shower or a bathtub. \n", "A bathroom is a room in the home or hotel for personal hygiene activities, generally containing a toilet, a sink (basin) and either a bathtub, a shower, or both. In some countries, the toilet is usually included in the bathroom, whereas other cultures consider this insanitary or impractical, and give that fixture a room of its own. The toilet may even be outside of the home in the case of pit latrines. It may also be a question of available space in the house whether the toilet is included in the bathroom or not. \n", "In Canadian English, public facilities are frequently called \"washrooms\", although usage varies regionally. The word \"toilet\" generally denotes the fixture itself rather than the room. The word \"washroom\" is rarely used to mean \"utility room\" or \"mud room\" as it is in some parts of the United States. \"Bathroom\" is generally used to refer to the room in a person's home that includes a bathtub or shower. In public athletic or aquatic facilities, showers are available in locker rooms.\n", "Generally, patients who are able to are encouraged to walk to the toilet or use a bedside commode as opposed to a urinal. The prolonged use of a urinal has been shown to lead to constipation or trouble urinating.\n", "Inability to access a bathroom when necessary has caused health issues such as urinary tract infections, kidney infections, and digestive problems which can later develop into severe health problems. Inadequate access to the use of a bathroom when required can lead to substantial problems for people who have prostate problems, going through menopause, or are menstruating for instance.\n", "In North American English the word \"bathroom\" may be used to mean any room containing a toilet, even a public toilet (although in the United States this is more commonly called a restroom and in Canada a washroom).\n" ]
why do prepaid visa cards ask for social security numbers?
Its a bank account. Just like a bank. You can't open an account at a brick and mortar bank without using a SSN. Same applies to a non-traditional account.
[ "Social Security numbers and cards are issued by the US Social Security Administration for tracking of Social Security taxes and benefits. They have become the \"de facto\" national identification number for federal and state taxation, private financial services, and identification with various companies. SSNs do not establish citizenship because they can also be issued to permanent residents as well as citizens. They typically can only be part of the establishment of a person's identity; a photo ID that verifies date of birth is also usually requested.\n", "In 2004 Congress passed The Intelligence Reform and Terrorism Prevention Act; parts of which mandated that the Social Security Administration redesign the Social Security Number (SSN) Card to prevent forgery. From April 2006 through August 2007, Social Security Administration (SSA) and Government Printing Office (GPO) employees were assigned to redesign the Social Security Number Card to the specifications of the Interagency Task Force created by the Commissioner of Social Security in consultation with the Secretary of Homeland Security.\n", "The Social Security number (SSN) and card are issued by the Social Security Administration. Almost all parents voluntarily apply for a Social Security number shortly after the birth of a child. In the absence of a national identity card (and concordant national identity number), the Social Security number has become the \"de facto\" national identifier for a large variety of purposes, both governmental and non-governmental.\n", "Prior to the 2011 randomization process, the first three digits or area number were assigned by geographical region. Prior to 1973, cards were issued in local Social Security offices around the country and the area number represented the office code where the card was issued. This did not necessarily have to be in the area where the applicant lived, since a person could apply for their card in any Social Security office. Beginning in 1973, when the SSA began assigning SSNs and issuing cards centrally from Baltimore, the area number was assigned based on the ZIP Code in the mailing address provided on the application for the original Social Security card. The applicant's mailing address did not have to be the same as their place of residence. Thus, the area number did not necessarily represent the state of residence of the applicant regardless of whether the card was issued prior to, or after, 1973.\n", "Social Security cards have federal jurisdiction but cannot verify identity. They verify only the match between a given name and a Social Security Number (SSN) and were intended only for use in complying with Social Security payroll tax laws. They now are used in a wider scope of activities, such as for obtaining credit and other regulated financial services in banking and investments.\n", "The Social Security Administration has suggested that, if asked to provide his or her Social Security number, a citizen should ask which law requires its use. In accordance with §7213 of the 9/11 Commission Implementation Act of 2004 and , the number of replacement Social Security cards per person is generally limited to three per calendar year and ten in a lifetime.\n", "Every citizen has an NIC number for activities such as paying taxes, opening a bank account, getting a utility connection (phone, cell phone, gas, electricity). However, since a majority of births in the country are not registered, and a large number of Pakistanis do not conduct any of the activities described above, most do not have ID cards. Obtaining an NIC card costs 100 rupees (US$1.66 - almost the average daily income), and this reduces the number of people who can afford it. In 2006, NADRA announced that it had issued 50 million CNIC (the C standing for Computerized) numbers, which is approximately one-third of the population. In June 2008, the federal government announced it would start issuing CNIC cards for free.\n" ]
How often are planets found?
Planets in our Solar System, or planets in general? In our Solar System, only two planets still considered planets have been found in recorded history, Uranus in 1781, and Neptune in 1864. If Planet Nine is real, odds are we'll find that within the next 10 years or so. Outside of our solar system, planets are found on a virtually daily basis, although usually only announced in batches of a few planets, a few dozen planets, or rarely a few hundred planets. For some planets just announced recently, a bunch of planets around nearby stars were announced only a week ago: [_URL_0_](_URL_0_)
[ "BULLET::::- 30 planets: On October 19, it was announced that 30 new planets were discovered, all were detected by radial velocity method. It is the most planets ever announced in a single day during the exoplanet era. October 2009 now holds the most planets discovered in a month, breaking the record set in June 2002 and August 2009, during which 17 planets were discovered.\n", "Sixteen candidate planets were discovered with orbital periods ranging from 0.6 to 4.2 days. Planets with orbital periods less than 1.2 days have not previously been detected, and have been dubbed \"ultra-short period planets\" (USPPs) by the search team. USPPs were discovered only around low-mass stars, suggesting that larger stars destroyed any planets orbiting so closely or that planets were unable to migrate as far inward around larger stars.\n", "BULLET::::- In January 2015, the number of confirmed Kepler planets exceeded 1000. At least two (Kepler-438b and Kepler-442b) of the discovered planets announced that month were likely rocky and in the habitable zone. Also in January 2015, NASA reported that five confirmed sub-earth-sized rocky exoplanets, all smaller than the planet Venus, were found orbiting the 11.2 billion year old star Kepler-444, making this star system, at 80% of the age of the universe, the oldest yet discovered.\n", "In 2005, the first planets around main-sequence stars that may be terrestrial were found: Gliese 876 d, has a mass 7 to 9 times that of Earth and an orbital period of just two Earth days. It orbits the red dwarf Gliese 876, 15 light years from Earth. OGLE-2005-BLG-390Lb, about 5.5 times the mass of Earth, orbits a star about 21,000 light years away in the constellation Scorpius.\n", "In 2010, the discovery of a third planet (47 UMa d) was made by using the Bayesian Kepler Periodogram. Using this model of this planetary system it was determined that it is 100,000 times more likely to have three planets than two planets. This discovery was announced by Debra Fischer and P.C. Gregory. This planet has an orbital period of 14,002 days or 38.33 years and a semi-major axis of 11.6 AU with a moderate eccentricity of 0.16. It would be the longest-period planet discovered by the radial velocity method, although longer-period planets had previously been discovered by direct imaging and pulsar timing.\n", "In 2018, a planet was discovered orbiting 40 Eridani A with a minimum mass of Earth masses. The planet has an orbit of 42 days, and lies considerably interior to the habitable zone, receiving 9 times more stellar flux than Earth, which is an even greater stellar flux amount than Mercury, the innermost planet in our solar system, on average receives from our Sun. It is one of the closest Super-Earths known, the closest discovered to date () within a multiple star system.\n", "Based on Kepler's findings, astronomer Seth Shostak estimated in 2011 that \"within a thousand light-years of Earth\", there are \"at least 30,000\" habitable planets. Also based on the findings, the Kepler team has estimated that there are \"at least 50 billion planets in the Milky Way\", of which \"at least 500 million\" are in the habitable zone. In March 2011, astronomers at NASA's Jet Propulsion Laboratory (JPL) reported that about \"1.4 to 2.7 percent\" of all Sun-like stars are expected to have Earth-size planets \"within the habitable zones of their stars\". This means there are \"two billion\" of these \"Earth analogs\" in the Milky Way alone. The JPL astronomers also noted that there are \"50 billion other galaxies\", potentially yielding more than one sextillion \"Earth analog\" planets if all galaxies have similar numbers of planets to the Milky Way.\n" ]
whats happening when a sneeze ‘gets stuck’ then just burns your nose and makes your eyes water.
Sneezes are a protective response to alert you to less than ideal breathing conditions and remove irritants/allergens from your nose. They’re triggered by the presence of irritants, but only a certain concentration, which is mediated by multiple nerve endings that generate “spikes” when they’re irritated. Once the number of spikes passes a certain threshold, you sneeze. Sometimes your nose will be irritated to the point of feeling like you have to sneeze, but there isn’t quite enough to push you over the threshold. So you “get stuck.”
[ "Sneezing typically occurs when foreign particles or sufficient external stimulants pass through the nasal hairs to reach the nasal mucosa. This triggers the release of histamines, which irritate the nerve cells in the nose, resulting in signals being sent to the brain to initiate the sneeze through the trigeminal nerve network. The brain then relates this initial signal, activates the pharyngeal and tracheal muscles and creates a large opening of the nasal and oral cavities, resulting in a powerful release of air and bioparticles. The powerful nature of a sneeze is attributed to its involvement of numerous organs of the upper body – it is a reflexive response involving the face, throat, and chest muscles.\n", "There is much debate about the true cause and mechanism of the sneezing fits brought about by the photic sneeze reflex. Sneezing occurs in response to irritation in the nasal cavity, which results in an afferent nerve fiber signal propagating through the ophthalmic and maxillary branches of the trigeminal nerve to the trigeminal nerve nuclei in the brainstem. The signal is interpreted in the trigeminal nerve nuclei, and an efferent nerve fiber signal goes to different parts of the body, such as mucous glands and the thoracic diaphragm, thus producing a sneeze. The most obvious difference between a normal sneeze and a photic sneeze is the stimulus: normal sneezes occur due to irritation in the nasal cavity, while the photic sneeze can result from a wide variety of stimuli. Some theories are below. There is also a genetic factor that increases the probability of photic sneeze reflex. The C allele on the rs10427255 SNP is particularly implicated in this although the mechanism is unknown by which this gene increases the probability of this response.\n", "A sneeze, or sternutation, is a semi-autonomous, convulsive expulsion of air from the lungs through the nose and mouth, usually caused by foreign particles irritating the nasal mucosa. A sneeze expels air forcibly from the mouth and nose in an explosive, spasmodic involuntary action resulting chiefly from irritation of the nasal mucous membrane. This action allows for mucus to escape through the nasal cavity. Sneezing is possibly linked to sudden exposure to bright light, sudden change (fall) in temperature, breeze of cold air, a particularly full stomach, or viral infection, and can lead to the spread of disease.\n", "Pepper is known to cause sneezing. Some sources say that piperine, a substance present in black pepper, irritates the nostrils, causing the sneezing. Few, if any, controlled studies have been carried out to answer the question.\n", "The sneeze reflex involves contraction of a number of different muscles and muscle groups throughout the body, typically including the eyelids. The common suggestion that it is impossible to sneeze with one's eyes open is, however, inaccurate. Other than irritating foreign particles, allergies or possible illness, another stimulus is sudden exposure to bright light – a condition known as photic sneeze reflex (PSR). Walking out of a dark building into sunshine may trigger PSR, or the ACHOO (autosomal dominant compulsive helio-ophthalmic outbursts of sneezing) syndrome as it's also called. The tendency to sneeze upon exposure to bright light is an autosomal dominant trait and affects 18-35% of the human population. A rarer trigger, observed in some individuals, is the fullness of the stomach immediately after a large meal. This is known as snatiation and is regarded as a medical disorder passed along genetically as an autosomal dominant trait.\n", "The function of sneezing is to expel mucus containing foreign particles or irritants and cleanse the nasal cavity. During a sneeze, the soft palate and palatine uvula depress while the back of the tongue elevates to partially close the passage to the mouth so that air ejected from the lungs may be expelled through the nose. Because the closing of the mouth is partial, a considerable amount of this air is usually also expelled from the mouth. The force and extent of the expulsion of the air through the nose varies.\n", "Uncontrollable fits of sneezing are common in patients under propofol sedation who undergo periocular or retrobulbar injection. A sneeze by a sedated patient often occurs upon insertion of a needle into or around their eye. The violent and uncontrollable movement of the head during a reflexive sneeze has potential to cause damage within the patient's eye if the needle is not removed before the sneeze occurs.\n" ]
What was Moscow's relationship with Ceausescu before the Romanian revolution? What role, if any, did they play in the revolution?
Ceausescu was something of a maverick in the Eastern Bloc, which is why your friend probably thinks that. Relations between Brezhnev/Andropov and Ceausescu were cool at best and outright frigid at worst, although they never reached the point of de facto breaking off relations, unlike in Beijing. He denounced the 1968 Soviet invasion of Czechoslovakia. He kept relations with both Israel and the PLO and helped push for peace between Israel and Egypt. He kept up cordial relations with the Chinese, much to the extreme irritation of the Kremlin-Ceausescu personally modeled his personality cult off of Mao Zedong and Kim Il Sung. He openly recognized West Germany and was the first Warsaw Pact country to independently invite a US President to visit. (Nixon would later use Ceaucescu as a conduit for backdoor negotiations with the Vietnamese and while in Bucharest, consulted with him on his desire to open relations with China.) He refused to endorse the invasion of Afghanistan, and participated in the 1984 Olympics in Los Angeles, which the Soviet Union boycotted. With that being said, however, I don't find it quite plausible. To be sure, if there is one intelligence service in the world which **nothing** should be put beyond, it's Russia's, no matter who rules in the Kremlin. Westerners often have a hard time grasping just how tactically skilled and utterly amoral they are. But Moscow wasn't able to react in places with a much larger KGB presence and more strategic value, like East Germany, in 1989, as none other than Vladimir Putin points out quite vividly. ("Moscow was silent.") I don't think they could have orchestrated a coup in a place like Romania where the security service in practice watched out for the KGB as much as anybody else. Moreover, the 1989 Romanian Revolution wasn't a reaction against just Ceausescu, in contrast to previous coup attempts such as that in 1984(I think), but against Communism as a whole. No matter how many Communists joined the opposition, the ideology was strongly discredited all over the Warsaw Pact by 1989. Gorbachev thought he could keep the genie in the bottle. But the Romanian Communists who joined the Revolution didn't, as seen by their subsequent economic and social liberalization policies.
[ "Satisfied with this appointment, the Soviet authorities were more conciliatory with Romania. On 13 March 1945 Moscow transferred the administration of Transylvania to Bucharest. A few months later, on 19 July 1945, Michael I was decorated with the Order of Victory, one of the most prestigious Soviet military orders. Still, the Sovietization of the kingdom was accelerated. The purge of \"fascist\" personalities continued while censorship was strengthened. A land reform was also implemented, causing a drop in production which ruined agricultural exports. The king, however, managed to temporarily prevent the establishment of People's Tribunals and the restoration of the death penalty.\n", "Establishing the republic became a matter of dispute. Despite the objections of Soviet commissar of foreign relations Chicherin who argued that the new establishment would only strengthen the position of Romanians towards Bessarabia and able to activate \"expansionist claims of Romanian chauvinism\", Kremlin launched a campaign to create the autonomy attracting to it Bessarabian refugees and Romanian political emigrants who lived in Moscow and the Ukrainian Socialist Soviet Republic.\n", "In 1917, the Triple Alliance armies occupied Bucharest and were advancing towards Iași. The Romanian Government decided to move the State Treasury to Russia in two transports, delegating Alexandru I. Lapedatu to accompany the second one that included cultural goods. He left Iași for Moscow on 28 July 1917, and stayed there until 19 December 1917, experiencing the arrival of the Bolshevik revolution.\n", "The Soviet Union regained the region in 1944, and the Red Army occupied Romania. By 1947, the Soviets had imposed a communist government in Bucharest, which was friendly and obedient towards Moscow. The Soviet occupation of Romania lasted until 1958. The Romanian communist regime did not openly raise the matter of Bessarabia or Northern Bukovina in its diplomatic relations with the Soviet Union. At least 100,000 people died in a post-war famine in Moldavia.\n", "The Soviet Union regained the region in 1944, and the Red Army occupied Romania. By 1947, the Soviets had imposed a communist government in Bucharest, which was friendly and obedient towards Moscow. The Soviet occupation of Romania lasted until 1958. The Romanian communist regime did not openly raise the matter of Bessarabia or Northern Bukovina in its diplomatic relations with the Soviet Union. At least 100,000 people died in a post-war famine in Moldavia.\n", "After World War I relations between Romania and Soviet Russia were tense. Since 1918 there were numerous bilateral meetings in Copenhagen, Warsaw, Genoa, and other locations but no consensus could be reached. The Soviets saw Bessarabia as an annexed province and considered the decision of union with Romania as imposed by the occupying Romanian Army. Moreover, historians from both countries intensely debated the treaty with the Soviet Rumcherod in 1918 that required withdrawal of the Romanian Army from Bessarabia but which both countries failed to respect. The legitimacy of the Sfatul Țării was also brought into question, although the only contested decision is the unification act.\n", "The October Revolution shook Russia and placed most of it under a Bolshevik government which had no intention of continuing with war against the Central Powers. Although the Romanian presence in Kiev was set back by the November Uprising and the January Rebellion, then dispersed by the anti-Entente Skoropadsky regime, Constantin Gh. Pietraru and a small force remained behind in the new Ukrainian People's Republic (UNR), where they signed up the last group of Romanian volunteers. Some of these efforts were hampered by a diplomatic tensions between the UNR and Romania. Ukrainian officials refused to either rally with the Entente or negotiate border treaties with Romania, but tacitly permitted Deleu, Bocu, Ghibu and other Transylvanian Romanians activists who worked against Austria-Hungary to work on UNR territory.\n" ]
why do roosters "cock-a-doodle-doo" in the morning?
They actually crow all day. Not just mornings.
[ "A rooster's crowing is a loud and sometimes shrill call and sends a territorial signal to other roosters. However, roosters may also crow in response to sudden disturbances within their surroundings. Hens cluck loudly after laying an egg, and also to call their chicks. Chickens also give different warning calls when they sense a predator approaching from the air or on the ground.\n", "The rooster is often portrayed as crowing at the break of dawn (\"cock-a-doodle-doo\"). However, while many roosters crow shortly after waking up, this idea is not exactly true. A rooster can and will crow at any time of the day. Some roosters are especially vociferous, crowing almost constantly, while others only crow a few times a day. These differences are dependent both upon the rooster's breed and individual personality. A rooster can often be seen sitting on fence posts or other objects, where he crows to proclaim his territory.\n", "BULLET::::- Roosters can be heard crowing as it begins to get lighter. In past centuries people believed the rooster controlled the rise of daylight and thus only crowed at this occasion. While roosters do indeed crow at dawn and therefore were often used as a prototypical alarm clock in past centuries, they can and will crow at any time of the day, not just in the morning. The idea that the rooster scares the darkness away led to its worship in various religious belief systems. In English the word \"cock-crow\" is a synonym for \"early morning\".\n", "\"Roosting\" is the action of perching aloft to sleep at day, which is done by both sexes. The rooster is polygamous, but cannot guard several nests of eggs at once. He guards the general area where his hens are nesting, and attacks other roosters that enter his territory. During the daytime, a rooster often sits on a high perch, usually off the ground, to serve as a lookout for his group (hence the term \"rooster\"). He sounds a distinctive alarm call if predators are nearby and will frequently crow to assert his territory.\n", "BULLET::::- Roosters usually sit on high perches, looking out for their group. When it spots danger it will crow loudly. This led people to portray roosters as people who crave attention and suffer from delusions of grandeur. The image of the high perched rooster is also prevalent in Christian traditions, where statues of cocks are often put on top of church steeples as a weather vane.\n", "Roosters have several other calls as well, and can cluck, similar to the hen. Roosters occasionally make a patterned series of clucks to attract hens to a source of food, the same way a mother hen does for her chicks.\n", "Roosters almost always start crowing before four months of age. Although it is possible for a hen to crow as well, crowing (together with hackles development) is one of the clearest signs of being a rooster.\n" ]
How Accurate/Bias was "Century of the Self"?
Not commenting directly on "Century of the Self", but Adam Curtis: Most all of his documentaries start with "This is a story about..." Keep that in mind. It's a story. It's a particular account of events too complex to explain otherwise. It may be compelling, even undeniable. It may be well researched, he may have universal consensus backing his thesis or smoking gun evidence to back it up. But it's just a story. I like Adam Curtis, I think he illuminates interesting aspects of our culture and recent history. I agree with him on a lot of his views. But I'd still endorse some healthy scepticism when watching his work. There's often a lot he overlooks or neglects (because if he included it, his "story" wouldn't run). (I really wanna make a documentary about Adam Curtis that starts with "This is a story about a documentary maker who believed the chaos of historical events could be explained in a simple way, and the key to doing so lay in understanding the ideologies of those involved...")
[ "The term of present bias was coined in the second half of the 20th century. In the 1930’s economic research started investigating time preferences. The findings led to the model of exponential discounting, thus time consistent discounting. However, later research led to the conclusion that time preferences were indeed not consistent, but inconsistent. In other words, people were found to prefer immediate advantages to future advantages in that their discount over a short period of time falls rapidly, while falling less the more the rewards are in the future. Therefore, people are biased towards the present. As a result, Phelps and Pollak introduced the quasi-hyperbolic model in 1968. In economics, present bias is therefore a model of discounting.\n", "On Metacritic, \"The Century of Self\" has been given a score of 68 out of 100 based on \"generally favorable reviews.\" Clashmusic.com, the online arm of \"Clash\" magazine, gave it a positive review and commented:\n", "Along these lines, \"The Century of the Self\" asks deeper questions about the roots and methods of consumerism and commodification and their implications. It also questions the modern way people see themselves, the attitudes to fashion, and superficiality.\n", "Cultural bias may also arise in historical scholarship, when the standards, assumptions and conventions of the historian's own era are anachronistically used to report and assess events of the past. This tendency is sometimes known as presentism, and is regarded by many historians as a fault to be avoided. Arthur Marwick has argued that \"a grasp of the fact that past societies are very different from our own, and ... very difficult to get to know\" is an essential and fundamental skill of the professional historian; and that \"anachronism is still one of the most obvious faults when the unqualified (those expert in other disciplines, perhaps) attempt to do history\".\n", "The Century of the Self is a 2002 British television documentary series by filmmaker Adam Curtis. It focuses on the work of psychoanalysts Sigmund Freud and Anna Freud, and PR consultant Edward Bernays. In episode one, Curtis says, \"This series is about how those in power have used Freud's theories to try and control the dangerous crowd in an age of mass democracy.\"\n", "People who promote conspiracy theories or apocalyptic future scenarios have cited the normalcy bias as a prime reason why others scoff at their pronouncements. For example, survivalists who fear that the U.S. will soon descend into totalitarianism cite normalcy bias as the reason why most Americans do not share their worries. Similarly, fundamentalist Christians use the normalcy bias to explain why others scoff at their beliefs about the \"End Time\". One fundamentalist website writes: \"May we not get blinded by the 'normalcy bias' but rather live with the knowledge that the Lord’s coming is near.\"\n", "BULLET::::7. People favour \"plausibility over accuracy\" in accounts of events and contexts (Currie & Brown, 2003; Brown, 2005; Abolafia, 2010): \"in an equivocal, postmodern world, infused with the politics of interpretation and conflicting interests and inhabited by people with multiple shifting identities, an obsession with accuracy seems fruitless, and not of much practical help, either\" (Weick, 1995: 61).\n" ]
how does my (i)phone know which sounds to let through in a phone call / face time?
There are several microphones. 1 intended to pick up your voice, the other(s) to pick up all the rest of the noise around you (lets call them noise microphones). The cell phone subtracts the sounds from the noise microphones from the sounds picked up by the voice microphone. The sounds that are left are just the voice.
[ "A typical phone call using a traditional phone is placed by picking the phone handset up off the base and holding the handset so that the hearing end is next to the user's ear and the speaking end is within range of the mouth. The caller then rotary dials or presses buttons for the phone number needed to complete the call, and the call is routed to the phone which has that number. The second phone makes a ringing noise to alert its owner, while the user of the first phone hears a ringing noise in its earpiece. If the second phone is picked up, then the operators of the two units are able to talk to one another through them. If the phone is not picked up, the operator of the first phone continues to hear a ringing noise until they hang up their own phone.\n", "Typically, the disconnect tone is a few cycles of the reorder or busy tone (e.g. in US), or between five and fifteen seconds of the Number Unobtainable tone (e.g. in UK). On some telephone exchanges in the UK, the following audio message is looped for fifteen seconds, interspersed with special information tones (SIT), to advise the remote party has hung up: \"(SIT) The other person has hung up\". On iPhones, the tone is 3 bursts of 425 Hz tones each lasting 0.2 seconds.\n", "BULLET::::2. \"The phone keeps track of the 'Network connection time,' which is 'the elapsed time from the moment [the user] connect[s] to [the] service provider's network to the moment [the user] end[s] the call by pressing [the end key].'\" The court used this as evidence that the phone performs logical and arithmetic operations when placing calls.\n", "A dial tone is a telephony signal sent by a telephone exchange or private branch exchange (PBX) to a terminating device, such as a telephone, when an off-hook condition is detected. It indicates that the exchange is working and is ready to initiate a telephone call. The tone stops when the first dialed digit is recognized. If no digits are forthcoming, the permanent signal procedure is invoked, often eliciting a special information tone and an intercept message, followed by the off-hook tone, requiring the caller to hang up and redial.\n", "In Australia, the number 1194 gives the speaking clock in all areas and from all providers. It is always the current time from where the call originates. A male voice, often known by Australians as George, says \"At the third stroke, it will be (hours) (minutes) and (seconds) seconds/precisely. (three beeps)\" e.g. \"At the third stroke, it will be three thirty three and forty seconds ... beep beep beep\". These are done in 10 second increments and the beep is 1 kHz. Originally there was only one stroke eg:”At the stroke, it will be...” etc.\n", "ToneLoc took advantage of the extended return codes available on US Robotics modems (e.g., ATX6) to detect dial tones upon dialing a number and to detect when a human answered the phone in addition to scanning for other modems. Detection of voice numbers sped up the scanning process by disconnecting upon detecting a human instead of timing out waiting for a modem carrier signal. The detection of a dial tone after dialing a number allowed for users to search for poorly secured extenders which could be used to divert calls through. \n", "Mobiles would originate calls by sending a burst of connect tone, to which the base station responded with a burst of seize tone. The mobile would then respond with its identification, consisting of its area code and last four digits of the phone number sent at 20 pulses per second, just as in inward dialing but with the addition of rudimentary parity checking. Digits are formed with a pulsetrain of alternating tones, either connect and silence (for odd digits) or connect and guard (for even digits). When the base station received the calling party's identification, it would send dialtone to the mobile. The user would then use the rotary dial, which would send the dialed digits as an alternating 10 pps pulse train (originally, directly formed by the rotary dial) of connect and guard tones.\n" ]
In the post red giant stage of a star, why do the outer layers drift into space and not collapse onto the white dwarf?
Two reasons. One is that as the star expands the surface gravity decreases and so it's easier for those outer layers to escape. Two is that it's not actually a "drifting" away in many cases but is in fact forceful; [stellar winds](_URL_0_) are generated which are blowing surface material away. We see this observationally but the physical processes that are all intertwined are quite complex.
[ "A third option is that a giant star passes close enough to the central massive black hole for the outer layers to be stripped away by tidal forces, after which the remaining core may become an EMRI. However, it is uncertain if the coupling between the core and outer layers of giant stars is strong enough for stripping to have a significant enough effect on the orbit of the core.\n", "When the core of a medium mass star contracts, it causes a release of energy that makes the envelope of the star expand. This continues until the star finally blows its outer layers off. The core of the star remains intact and becomes a white dwarf. The white dwarf is surrounded by an expanding shell of gas in an object known as a planetary nebula. Planetary nebula seem to mark the transition of a medium mass star from red giant to white dwarf. X-ray images reveal clouds of multimillion degree gas that have been compressed and heated by the fast stellar wind. Eventually the central star collapses to form a white dwarf. For a billion or so years after a star collapses to form a white dwarf, it is \"white\" hot with surface temperatures of ~20,000 K.\n", "The evolved red giant star is losing mass, since radiation pressure overcomes the low gravity on the surface. The outflow of matter is captured by the gravitational field of the white dwarf and falls on its surface in the end. At least during the active phase an accretion disk forms around the white dwarf.\n", "White dwarfs resist gravitational collapse primarily through electron degeneracy pressure (compare main sequence stars, which resist collapse through thermal pressure). The Chandrasekhar limit is the mass above which electron degeneracy pressure in the star's core is insufficient to balance the star's own gravitational self-attraction. Consequently, a white dwarf with a mass greater than the limit is subject to further gravitational collapse, evolving into a different type of stellar remnant, such as a neutron star or black hole. Those with masses under the limit remain stable as white dwarfs.\n", "WD 1145+047 b is currently being vaporized by its star because of its extreme proximity to it. White dwarfs are usually the size of the Earth, and have half as much mass as they did during the main sequence. Due to this and the searing hot temperature of the stellar remnant, rocky minerals are being vaporized off the surface of this object, into orbit around the star, which is responsible for a hot dusty disk that was observed around its host star. It is likely that WD 1145+017 b is bound to disintegrate in the future (around 100–200 million years from now) due to further vaporization and ablation. The minor planet is likely being pelted by several smaller objects of up to , as it is likely not just a single object orbiting the white dwarf star, but likely several planetesimals, which is probably responsible for some of the variations in the light curve data. The smaller objects can also throw debris into orbit upon impact, which may also be responsible for the variations.\n", "If the white dwarf's mass increases above the Chandrasekhar limit, which is for a white dwarf composed chiefly of carbon, oxygen, neon, and/or magnesium, then electron degeneracy pressure fails due to electron capture and the star collapses. Depending upon the chemical composition and pre-collapse temperature in the center, this will lead either to collapse into a neutron star or runaway ignition of carbon and oxygen. Heavier elements favor continued core collapse, because they require a higher temperature to ignite, because electron capture onto these elements and their fusion products is easier; higher core temperatures favor runaway nuclear reaction, which halts core collapse and leads to a Type Ia supernova. These supernovae may be many times brighter than the Type II supernova marking the death of a massive star, even though the latter has the greater total energy release. This instability to collapse means that no white dwarf more massive than approximately can exist (with a possible minor exception for very rapidly spinning white dwarfs, whose centrifugal force due to rotation partially counteracts the weight of their matter). Mass transfer in a binary system may cause an initially stable white dwarf to surpass the Chandrasekhar limit.\n", "As a star's core shrinks, the intensity of radiation from that surface increases, creating such radiation pressure on the outer shell of gas that it will push those layers away, forming a planetary nebula. If what remains after the outer atmosphere has been shed is less than 1.4 , it shrinks to a relatively tiny object about the size of Earth, known as a white dwarf. White dwarfs lack the mass for further gravitational compression to take place. The electron-degenerate matter inside a white dwarf is no longer a plasma, even though stars are generally referred to as being spheres of plasma. Eventually, white dwarfs fade into black dwarfs over a very long period of time.\n" ]
why can't facebook, twitter and instagram just shut down bot accounts?
The problem is how do you actually figure out that a twitter user is a bot. You can use Machine Learning with certain features (like tweet sentiment) to analyze the data, but there's the possibility of false positives. Also, it's an arms race. Once they figure out what you're looking for all the bot creator needs to do is to change the bot themselves. You can take a look at [this pdf](_URL_0_) if you're interested in the details of figuring out whether a twitter user is a bot or a human. My ML professor participated in the DARPA Twitter Bot challenge and he said there were a lot of arguments on his team because of it, and at times he felt that it was going to rip the team apart.
[ "Using social bots is against the terms of service of many platforms, especially Twitter and Instagram. However, a certain degree of automation is of course intended by making social media APIs available. Many users, especially businesses still automate their Instagram activity in order to gain real followers rather than buying fake ones. This is commonly done through third-party social automation companies.\n", "BugMeNot allows users of their service to add new accounts for sites with free registration. It also encourages users to use disposable email address services to create such accounts. However, it does not allow them to add accounts for pay websites, as this could potentially lead to credit card fraud. BugMeNot also claims to remove accounts for any website requesting that they do not provide accounts for non-registered users.\n", "Due to the strict Internet censorship policy on microblogging enacted by the CPC government, a number of Chinese microbloggers choose to make posts that contain \"sensitive contents\" on Twitter. Although Twitter has been blocked in China since 2009, most Twitter users who reside in China can access the Twitter website using a proxy. More information can be found on List of websites blocked in China.\n", "A Twitter bot is a type of bot software that controls a Twitter account via the Twitter API. The bot software may autonomously perform actions such as tweeting, re-tweeting, liking, following, unfollowing, or direct messaging other accounts. The automation of Twitter accounts is governed by a set of automation rules that outline proper and improper uses of automation. Proper usage includes broadcasting helpful information, automatically generating interesting or creative content, and automatically replying to users via direct message. Improper usage includes circumventing API rate limits, violating user privacy, or spamming.\n", "Under the Terms of Service that Twitter requires its users to agree to, Twitter retains the right to temporarily or permanently suspend user accounts based on a violation of the agreement. One such example took place on December 18, 2017, when it banned the accounts belonging to Paul Golding, Jayda Fransen, Britain First, and the Traditionalist Worker Party.\n", "In addition to content-generating bots, users can purchase followers, favorites, retweets and comments on various websites that cater to expanding a user's image through accumulation of followers. With more followers, users' profiles gain more attention, thus increasing their popularity. Generating Web traffic is a valuable commodity for both individuals and businesses because it indicates notability. With Twitterbots, users are able to create the illusion of \"buzz\" on their site by obtaining followers from services such as Swenzy and underground suppliers who operate bot farms or click farms. The companies that facilitate this service create fake Twitter accounts that follow a number of people, some of these Twitter accounts may even post fake tweets to make it seem like they are real. This practice of obtaining mass amounts of twitterbots as followers is not permitted on Twitter. The emphasis on followers and likes as a measure of social capital has urged people to extend their circle to weak and latent ties to promote the idea of popularity for celebrities, politicians, musicians, public figures, and companies alike. According to \"The New York Times\", bots amass significant influence and have been noted to sway elections, influence the stock market, public appeal, and attack governments.\n", "As well as frustrating other users, the misuse of hashtags can lead to account suspensions. Twitter warns that adding hashtags to unrelated tweets, or repeated use of the same hashtag without adding to a conversation, could cause an account to be filtered from search, or suspended.\n" ]
why do higher impedance speakers yield better sound quality?
The amplifier has an output impedance and the speakers have an input impedance. The signal itself is a time-varying voltage which is divided between the two impedances according to their ratio. To maximize the amount of power in the speakers, you want as low an output impedance on the amplifier and as high an input impedance on the speakers as possible. Note that this is the same for *any* sort of amplification process. You want high impedance on the load because that's where you want most of the signal to be dissipated. If you have a low impedance on the load, most of the signal will be dissipated (uselessly) elsewhere.
[ "Speaker sensitivity is measured and rated on the assumption of a fixed amplifier output voltage because audio amplifiers tend to behave like voltage sources. Sensitivity can be a misleading metric due to differences in speaker impedance between differently designed speakers. A speaker with a higher impedance may have lower measured sensitivity and thus appear to be less efficient than a speaker with a lower impedance even though their efficiencies are actually similar. Speaker efficiency is a metric that only measures the actual percentage of electrical power that the speaker converts to acoustic power and is sometimes a more appropriate metric to use when investigating ways to achieve a given acoustic power from a speaker.\n", "Using an under-rated, long, skinny, oxidized high impedance speaker cable will drastically reduce the damping properties of the entire audio system, which is why in many high-end audio systems, the amplifier is located as close to the speakers as possible.\n", "Large amounts of damping of the loudspeaker is not necessarily better, for example a mere 0.35 dB difference in real-life results between a high (100) and medium (20) Damping Factor. Some engineers, including Nelson Pass claim loudspeakers can sound \"better\" with lower electrical damping. A lower damping factor helps to enhance the bass response of the loudspeaker by several decibels (where the impedance of the speaker would be at its maximum), which is useful if only a single speaker is used for the entire audio range. Therefore, some amplifiers, in particular vintage amplifiers from the 1950s, '60s and '70s, feature controls for varying the damping factor. While such bass \"enhancement\" may be pleasing to some enthusiasts, it nonetheless represents a distortion of the input signal.\n", "This product was important because it was the first loudspeaker system which, due to its particular crossover, showed a flat impedance curve (except for the resonance peak) with a consequent easier work for the amplifier and linear frequency response.\n", "Vereker also believed that a well-designed amplifier must be stable at all times when driving real-life loads, which are different from those achieved in lab conditions because loudspeakers' impedances vary with frequency. The inherent compromise between the pursuits for stability and sound quality means that Naim's power amplifiers are designed to work optimally with its own moderately priced speaker cable, and its predecessor . Product manuals warn users against using \"high-definition wire or any other special cable between amplifier and loudspeaker\". Whilst other manufacturers habitually employ Zobel networks (or an output filter which enhances amplifiers' stability) to protect against use with speakers and or cables with very high-capacitance, Naim amplifiers routinely omit these filters because of their adverse effect on sound quality. The design decision was made to use a suitable length of speaker cable (a minimum of 3.5m, with 5m being optimal) to provide the effective inductance.\n", "The design of speaker crossover networks and other electro-mechanical properties may result in a speaker with a very uneven impedance curve, for a nominal 8 Ω speaker, being as low as 6 Ω at some places and as high as 30–50 Ω elsewhere in the curve. An amplifier with little or no negative feedback will always perform poorly when faced with a speaker where little attention was paid to the impedance curve.\n", "Thus conventional speaker output, or the fidelity of the device, is distorted by physical limitations inherent in its design. These distortions have long been the limiting factor in commercial reproduction of strong high frequencies. To a lesser extent square wave characteristics are also problematic; the reproduction of square waves most stress a speaker cone.\n" ]
why did teachers always tell us to remove hats/caps when we enter inside a building? what does this signify?
It goes back to olden times when a knight would remove his face gear. It's simply a sign of respect to remove head pieces when in someone else's "home", and as well as during the Pledge of Allegiance.
[ "Wearing coats, boots or other outer garments inside someone’s home is often frowned upon as well. Sitting down to eat at table wearing a hat or coat etc. is even worse. Also one should remove one's hat when showing deference. Removing one's hat is also a form of respectful greeting: the origin of this is that knights were expected to remove their helmets when meeting their king; not doing so would be a sign of mistrust and hostility.\n", "Hats are a feature where users who belong to an company, project, or organization may choose to wear a \"hat\", indicating that they are speaking on behalf of the organization. This allows users to fluidly move between talking in an official capacity to talking as themselves without changing accounts. Specially colored red hats are worn by members of the Lobsters community that upkeep the site, marked as \"Sysop\".\n", "The members wear outrageous headwear during their performances. This is explained as an act of rebellion, remembering their slave ancestors, who had to work bare-headed in the sun. According to N.W.H., hats are a symbol for resistance and revolution since their hatless ancestors were too tired from working all day in the sun to revolt. This is a typical example of the bizarre logic the group uses to explain the deeper meanings behind their otherwise crude and base music and images.\n", "It is often claimed that undergraduates by custom do not wear their caps (or even that they can be fined for doing so). This is incorrect. Outdoors, caps may be worn, but it is customary to touch or raise one's cap as a salute to senior university or college officers. Like all other male members of the university (including graduates) other than the Chancellor, Vice-Chancellor and Proctors, male undergraduates must remove their caps during university ceremonies indoors.\n", "A form of a black hat known as a square cap (also mortarboard) is worn or carried. Properly, it is worn outdoors and carried indoors, except by people acting in an official capacity who customarily continue to wear it indoors. Although in practice few people wear (or even carry) a cap nowadays, they are nominally still required for graduates at the university; caps ceased to be compulsory for undergraduates in 1943 due to a shortage during the Second World War, and, after bringing them back for degree ceremonies in the Senate House only, were finally made entirely optional for undergraduates in 1953, though they are still not permitted to wear any other head covering with a gown.\n", "Students wearing granny dresses to school were suspended or sent home. In Oakland in 1965, girls were sent home for wearing granny dresses. In Kansas City, Missouri, a mother wore her own granny dress to school in an attempt to convince the principal to allow her daughter to wear one. In Trumansburg, New York, in 1966, three sisters were suspended from school for wearing the dress. The school's attorney claimed that both safety and possible class disruption were the reasons the dress was banned. The principal of the school felt that there was a danger of tripping on stairs because of the length of the dresses. Laura M. Lorraine, dean of the Analy Union High School also thought the length of the dresses made them difficult for walking up stairs. The school attorney felt that granny dresses were \"extreme\" and may encourage students to adopt other extreme forms of dressing. In some cases, school authorities just stated that it wasn't \"suitable school attire.\" In 1966, a \"Dear Abby\" column featured a letter from a girl who was sent to the principal's office for wearing a granny dress.\n", "Men wear a mortarboard (also known as a \"square\" or trencher cap) [h1], which is not worn indoors, except by the Chancellor, Vice-Chancellor and Proctors. When meeting the Vice-Chancellor, Proctors, or other senior official of the university in the street, it is traditional for a man to touch or raise his cap. In practice few people wear their caps nowadays, and instead carry their caps on occasions where caps are required.\n" ]
Would having a more efficient/ faster brain affect our perception of time?
I am by no means an expert, but I know (both firsthand and from [documented sources](_URL_0_) ) that your perception of time can be affected by life-threatening situations. These situations *probably (and I'm guessing here)* do something like speeding up your neuronal activity and cause your neural "clock speed" to increase. This causes things around you to move slowly from your point of view. Perhaps someone in this field can expand further on this, but it's an interesting subject and I wish I knew more about it to properly answer your question.
[ "The neural efficiency hypothesis postulates that more intelligent individuals display less activation in the brain during cognitive tasks, as measured by Glucose metabolism. A small sample of participants (N=8) displayed negative correlations between intelligence and absolute regional metabolic rates ranging from -0.48 to -0.84, as measured by PET scans, indicating that brighter individuals were more effective processors of information, as they use less energy. According to an extensive review by Neubauer & Fink a large number of studies (N=27) have confirmed this finding using methods such as PET scans, EEG and fMRI.\n", "In this case, age is looked at as a factor that alters and degrades memory efficiency and abilities over time. Age-related memory problems become more persistent in the elderly years, and one's ability to recall previously encoded stimuli without cues or context is no longer optimal. However, verbal or visual stimuli can be recognized at the same level of efficiency over the course of a lifetime. Craik and his colleagues found physiological evidence for this cognitive degradation through their research into the brains of elderly participants. Specifically, they discovered that there is a reduction in frontal activity. Still, there is an increased level of activity in the left prefrontal cortex when older adults undergo some nonverbal tasks of retrieval when compared to younger individuals. Moreover, the presence of increased left prefrontal cortex activity is only found in tasks revolving retrieval but there is still a reduction when performing encoding tasks.\n", "Speed of processing is another theory that has been raised to explain working memory deficits. As a result of various studies he has completed examining this topic, Salthouse argues that as we age our speed of processing information decreases significantly. It is this decrease in processing speed that is then responsible for our inability to use working memory efficiently as we age. The younger persons brain is able to obtain and process information at a quicker rate which allows for subsequent integration and manipulation needed to complete the cognitive task at hand. As this processing slows, cognitive tasks that rely on quick processing speed then become more difficult.\n", "fMRI and EEG studies have revealed that task difficulty is an important factor affecting neural efficiency. More intelligent individuals display neural efficiency only when faced with tasks of subjectively easy to moderate difficulty, while no neural efficiency can be found during difficult tasks. In fact, more able individuals appear to invest more cortical resources in tasks of high difficulty. This appears to be especially true for the Prefrontal Cortex, as individuals with higher intelligence displayed increased activation of this area during difficult tasks compared to individuals with lower intelligence. It has been proposed that the main reason for the neural efficiency phenomenon could be that individuals with high intelligence are better at blocking out interfering information than individuals with low intelligence.\n", "A study by Khan et al. (2008) examined the influence of cognitive load (low vs. high) on time-based prospective memory. The findings implied that time-based prospective memory is severely affected when cognitive load is high. The study attributed the poor performance on time-based tasks as a result of dividing attentional resources into actively monitoring time, self-initiating the response at the appropriate time and the ongoing task. Humans have limited attentional capacity, and therefore high cognitive load affects monitoring of time and consequently time-based prospective memory performance negatively. Numerous aspects of daily life depend on time-based prospective memory, ranging from daily activities such as remembering what time to meet a friend, to more important tasks such as remembering what time to take medication.\n", "Psychologists have found that the subjective perception of the passing of time tends to speed up with increasing age in humans. This often causes people to increasingly underestimate a given interval of time as they age. This fact can likely be attributed to a variety of age-related changes in the aging brain, such as the lowering in dopaminergic levels with older age; however, the details are still being debated. In an experimental study involving a group of subjects aged between 19 and 24 and a group between 60 and 80, the participants' abilities to estimate 3 minutes of time were compared. The study found that an average of 3 minutes and 3 seconds passed when participants in the younger group estimated that 3 minutes had passed, whereas the older group's estimate for when 3 minutes had passed came after an average of 3 minutes and 40 seconds.\n", "The neural efficiency hypothesis is the phenomenon where smarter individuals show lower (more efficient) brain activation than less bright individuals on cognitive tests of low to moderate difficulty. For tasks of higher difficulty, however, smarter individuals show higher brain activation.\n" ]
Are you born allergic to things or do you get it later on?
No. You are not born allergic to anything. You need to be sensitized to the allergen first, and you aren't really fully capable of that until several months after birth. You can, however, be born with a predisposition to becoming allergic to things (not specific things, just "things" in general).
[ "An allergic reaction can be caused by any form of direct contact with the allergen—consuming food or drink one is sensitive to (ingestion), breathing in pollen, perfume or pet dander (inhalation), or brushing a body part against an allergy-causing plant (direct contact). Other common causes of serious allergy are wasp, fire ant and bee stings, penicillin, and latex. An extremely serious form of an allergic reaction is called anaphylaxis. One form of treatment is the administration of sterile epinephrine to the person experiencing anaphylaxis, which suppresses the body's overreaction to the allergen, and allows for the patient to be transported to a medical facility.\n", "The symptoms of allergic contact may persist for as long as one month before resolving completely. Once an individual has developed a skin reaction to a certain substance it is most likely that they will have it for the rest of their life, and the symptoms will reappear when in contact with the allergen.\n", "A wide variety of foods can cause allergic reactions, but 90% of allergic responses to foods are caused by cow's milk, soy, eggs, wheat, peanuts, tree nuts, fish, and shellfish. Other food allergies, affecting less than 1 person per 10,000 population, may be considered \"rare\". The use of hydrolysed milk baby formula versus standard milk baby formula does not appear to change the risk.\n", "Some susceptible individuals have experienced an allergic reaction after touching \"Suillus americanus\". The symptoms of allergic contact dermatitis generally develop one to two days after initial contact, persist for roughly a week, then disappear without treatment. Cooking the fruit bodies inactivates the responsible allergens.\n", "Some people are allergic to coriander leaves or seeds, having symptoms similar to those of other food allergies. In one study, 32% of pin-prick tests in children and 23% in adults were positive for coriander and other members of the family Apiaceae, including caraway, fennel, and celery. The allergic symptoms may be minor or life-threatening.\n", "Many substances can trigger an allergic reaction. Common triggers of a reaction include foods, likes nuts, eggs, milk, gluten, fruit and vegetables; insect bites from bees or wasps (often a severe response occurs); environmental factors such as pollen, dust, mold, plants like grass or trees, animal dander; medications or chemicals. Some people experience an allergic response to cold or hot temperatures outside, jewelry or sunlight. \n", "While the most obvious route for an allergic exposure is unintentional ingestion, some reactions are possible through external exposure. Peanut allergies are much more common in infants who had oozing and crusted skin rashes as infants. Sensitive children may react via ingestion, inhalation, or skin contact to peanut allergens which have persistence in the environment, possibly lasting over months.\n" ]
Will the other side of the moon ever be facing earth?
The Moon is [tidally locked](_URL_0_), meaning it's orbital period equals its rotational period. Thus, the answer to your question is that the same side of the Moon will *always* face the Earth. Viewing the Moon from Earth, we will never see the other side. Something very dramatic would have to happen to change that.
[ "As a result of the Moon's synchronous rotation, one side of the Moon (the \"near side\") is permanently turned towards Earth, and the other side, the \"far side\", mostly cannot be seen from Earth. This means, conversely, that Earth can be seen only from the near side of the Moon and would always be invisible from the far side.\n", "As a result of the Moon's synchronous rotation, one side of the Moon (the \"near side\") is permanently turned towards Earth, and the other side, the \"far side\", mostly cannot be seen from Earth. This means, conversely, that Earth can be seen only from the near side of the Moon and would always be invisible from the far side.\n", "The near side of the Moon is the lunar hemisphere that is permanently turned towards Earth, whereas the opposite side is the far side. Only one side of the Moon is visible from Earth because the Moon rotates on its axis at the same rate that the Moon orbits the Earth – a situation known as synchronous rotation, or tidal locking.\n", "The near side of the Moon, containing these maria that make up the man, is always facing Earth. This is due to a tidal locking or synchronous orbit. Thought to have occurred because of the gravitational forces partially caused by the Moon's oblong shape, its rotation has slowed to the point where it rotates exactly once on each trip around the Earth. This causes the near side of the Moon to always turn its face toward Earth.\n", "On the other side as the Moon from the Earth headed towards the left at New Zealand, as the umbral path was outside the South Pole and over the Prime Meridian to the Peninsula, the Moon from the Earth was seen as it was going on bottom, then on the right and on top in the peninsular portion though the Earth rotates to the east as it was north of the South Pole at the Prime Meridian, the rest of the world saw the Moon from the Earth headed towards the left.\n", "The Moon keeps one hemisphere of itself facing the Earth, due to tidal locking. Therefore, the first view of the far side of the Moon was not possible until the Soviet probe Luna 3 reached the Moon on October 7th, 1959 and further lunar exploration by the U.S. and the Soviet Union. However, this simple picture is only approximately true: over time, slightly \"more\" than half (about 59%) of the Moon's surface is seen from Earth due to libration.\n", "The Moon is in synchronous rotation as it orbits Earth; it rotates about its axis in about the same time it takes to orbit Earth. This results in it always keeping nearly the same face turned towards Earth. However, because of the effect of libration, about 59% of the Moon's surface can actually be seen from Earth. The side of the Moon that faces Earth is called the near side, and the opposite the far side. The far side is often inaccurately called the \"dark side\", but it is in fact illuminated as often as the near side: once every 29.5 Earth days. During new moon, the near side is dark.\n" ]
Question about image scaling
It's a bit more complicated than that. [Here's a very good explanation of how it's done.](_URL_0_)
[ "Image scaling can be interpreted as a form of image resampling or image reconstruction from the view of the Nyquist sampling theorem. According to the theorem, downsampling to a smaller image from a higher-resolution original can only be carried out after applying a suitable 2D anti-aliasing filter to prevent aliasing artifacts. The image is reduced to the information that can be carried by the smaller image.\n", "In computer vision and biological vision, scaling transformations arise because of the perspective image mapping and because of objects having different physical size in the world. In these areas, scale invariance refers to local image descriptors or visual representations of the image data that remain invariant when the local scale in the image domain is changed. \n", "Image scaling is used in, among other applications, web browsers, image editors, image and file viewers, software magnifiers, digital zoom, the process of generating thumbnail images and when outputting images through screens or printers.\n", "In computer graphics and digital imaging, image scaling refers to the resizing of a digital image. In video technology, the magnification of digital material is known as upscaling or resolution enhancement.\n", "By taking into account the edges in an image, this scaling algorithm reduces artifacts common to other image scaling algorithms. For example, staircase artifacts on diagonal lines and curves are eliminated.\n", "A scaling in the most general sense is any affine transformation with a diagonalizable matrix. It includes the case that the three directions of scaling are not perpendicular. It includes also the case that one or more scale factors are equal to zero (projection), and the case of one or more negative scale factors. The latter corresponds to a combination of scaling proper and a kind of reflection: along lines in a particular direction we take the reflection in the point of intersection with a plane that need not be perpendicular; therefore it is more general than ordinary reflection in the plane.\n", "When scaling a vector graphic image, the graphic primitives that make up the image can be scaled using geometric transformations, with no loss of image quality. When scaling a raster graphics image, a new image with a higher or lower number of pixels must be generated. In the case of decreasing the pixel number (scaling down) this usually results in a visible quality loss. From the standpoint of digital signal processing, the scaling of raster graphics is a two-dimensional example of sample-rate conversion, the conversion of a discrete signal from a sampling rate (in this case the local sampling rate) to another.\n" ]
how exactly do organisms evolve and adapt to their environment?
They don't. There's a bit of variation between individuals of a generation, and traits that prove beneficial have a better chance of passing on than traits that don't. If 95% of frogs of a species are green and 5% are red, so long as green is the preferred camouflage color, the green frogs will reproduce more on account of the red ones getting eaten. If, say, an environmental event occurs that causes red to be a better form of camouflage, you'll see more and more red frogs surviving to breed every generation, as the green frogs try to hide, fail, and get eaten. It might take generations upon generations, but eventually you might see a day where most of the frogs are red.
[ "Habitats and biota do frequently change. Therefore, it follows that the process of adaptation is never finally complete. Over time, it may happen that the environment changes little, and the species comes to fit its surroundings better and better. On the other hand, it may happen that changes in the environment occur relatively rapidly, and then the species becomes less and less well adapted. Seen like this, adaptation is a genetic \"tracking process\", which goes on all the time to some extent, but especially when the population cannot or does not move to another, less hostile area. Given enough genetic change, as well as specific demographic conditions, an adaptation may be enough to bring a population back from the brink of extinction in a process called evolutionary rescue. Adaptation does affect, to some extent, every species in a particular ecosystem.\n", "Populations change through the process of evolution. Each individual in a population has a unique role in their particular environment. This role, commonly known as an ecological niche, is simply how an organism lives in an environment in relation to others. Over successive generations, the organism must adapt to their surrounding conditions in order to develop their niche. An organism's niche will evolve as changes in the external environment occur. The most successful species in nature are those that are able to use adaptive behaviors to build on prior knowledge, thereby increasing their overall knowledge bank. In turn, this will improve their overall survival and reproductive success.\n", "Evolution influences every aspect of the form and behaviour of organisms. Most prominent are the specific behavioural and physical adaptations that are the outcome of natural selection. These adaptations increase fitness by aiding activities such as finding food, avoiding predators or attracting mates. Organisms can also respond to selection by cooperating with each other, usually by aiding their relatives or engaging in mutually beneficial symbiosis. In the longer term, evolution produces new species through splitting ancestral populations of organisms into new groups that cannot or will not interbreed.\n", "Over many generations, the genomes of organisms can change significantly, resulting in evolution. In the process called adaptation, selection for beneficial mutations can cause a species to evolve into forms better able to survive in their environment. New species are formed through the process of speciation, often caused by geographical separations that prevent populations from exchanging genes with each other.\n", "Along with standard biological evolution, life and planet co-evolve. Since the best adaptations are those that suit the ecological niche that the organism lives in, the physical and chemical characteristics of the environment drive the evolution of life by natural selection, but the opposite can also be true: with every advent of evolution, the environment changes.\n", "In the course of development, organisms help shape their internal and external environment, and in this way, influence their own development. Organisms also construct developmental environments for their offspring through various forms of extra-genetic inheritance.\n", "While environmental factors are significantly responsible for evolutionary change, they act merely as agents for natural selection. Some of the changes develop through interactions with pathogens. Change is inherently brought about via phenomena at the genetic level – mutations, chromosomal rearrangements and epigenetic changes. While the general types of mutations hold true across the living world, in plants, some other mechanisms have been implicated as highly significant.\n" ]
Why are galaxies the colour they are?
The colors gives insight in the composition of the star population of the galaxy. Blue regions are composed mostly of young hot stars, while red regions are older, cooler stars. You will often also see smaller, pink spots. These are huge clouds of hydrogen, the color stems from the characteristic emission of hydrogen gas, which you can see in the [balmer series](_URL_0_). The HII line is the bright red emission on the right side of the spectrum. When you look at galaxies in other wavelengths than the visible light, you can get a lot mor information about the galaxies composition, like radio or gamma emission. But those do not contribute to the visible appearance of the galaxy. It should also be noted, that most galaxies are not visible to the naked eye at all. They are so far away, that the expansion of the universe stretched their light into and beyond the infrared spectrum, and can only be seen by specialized equipment. Images of galaxies are then "converted" to what a human might see if he were close enough. But it's always an artists impression.
[ "In astronomy, a blue compact dwarf galaxy (BCD galaxy) is a small galaxy which contains large clusters of young, hot, massive stars. These stars, the brightest of which are blue, cause the galaxy itself to appear blue in colour. Most BCD galaxies are also classified as dwarf irregular galaxies or as dwarf lenticular galaxies. Because they are composed of star clusters, BCD galaxies lack a uniform shape. They consume gas intensely, which causes their stars to become very violent when forming.\n", "Some spiral galaxies have central bar-shaped structures composed of stars. These galaxies are called 'barred spirals' and have been investigated by Galaxy Zoo in several studies. It is unclear why some spiral galaxies have bars and some do not. Galaxy Zoo research has shown that red spirals are about twice as likely to host bars as blue spirals. These colours are significant. Blue galaxies get their hue from the hot young stars they contain, implying that they are forming stars in large numbers. In red galaxies, this star formation has stopped, leaving behind the cooler, long-lived stars that give them their red colour. Karen Masters, a scientist involved in the studies, stated: \"For some time data have hinted that spirals with more old stars are more likely to have bars, but with such a large number of bar classifications we're much more confident about our results. It's not yet clear whether the bars are some side effect of an external process that turns spiral galaxies red, or if they alone can cause this transformation.\"\n", "Mainstream astronomical theory before Galaxy Zoo held that elliptical (or 'early type') galaxies were red in color and spiral (or 'late type') galaxies were blue in color: several papers published as a result of Galaxy Zoo have proved otherwise. A population of blue ellipticals was found. These are galaxies which have changed their shape from spiral to oval, but still have young stars in them. Indeed, Galaxy Zoo came about through Schawinski's searching for blue elliptical galaxies, as near the end of 2006, he had spent most of his waking hours trying to find these rare galaxies. Blueness in galaxies means that new stars are forming. However ellipticals are almost always red, indicating that they are full of old and dead stars. Thus, blue ellipticals are paradoxical, but give clues to star-formation in different types of galaxies.\n", "The colors of the galaxy are peculiar. In NGC 4676A a core with some dark markings is surrounded by a bluish white remnant of spiral arms. The tail is unusual, starting out blue and terminating in a more yellowish color, despite the fact that the beginning of each arm in virtually every spiral galaxy starts yellow and terminates in a bluish color. NGC 4676B has a yellowish core and two arcs; arm remnants underneath are bluish as well.\n", "It is thought that Red Spirals are galaxies in the process of transition from young to old. They are more massive than blue spirals and are found on the outskirts of large clusters of galaxies. Chris Lintott stated: \"We think what we’re seeing is galaxies that have been gently strangled, so to speak, where somehow the gas supply for star formation has been cut off, but that they’ve been strangled so gently that the arms are still there.\" The cause might be the Red Spiral's gentle interaction with a galaxy cluster. He further explained: \"The kind of thing we’re imagining [is that] as the galaxy moves into a denser environment, there’s lot of gas in clusters as well as galaxies, and it’s possible the gas from the galaxy just gets stripped off by the denser medium it’s plowing into.\"\n", "Scientists hypothesize that many peculiar galaxies are formed by the collision of two or more galaxies. As such, peculiar galaxies tend to host more active galactic nuclei than normal galaxies, indicating that they contain supermassive black holes. Many peculiar galaxies experience starbursts, or episodes of rapid star formation, due to the galaxies merging. The periods of elevated star formation and the luminosity resulting from active galactic nuclei cause peculiar galaxies to be slightly bluer in color than other galaxies.\n", "Blue compact dwarf galaxies are blue in appearance because of the large cluster of hot massive stars. The galaxies are relatively dim and appear to be irregular in shape. They are typically chaotic in appearance.\n" ]
- why wouldn't it be possible to set up a solar heated steam pipeline that also removes salt, to get water to rural parts of africa?
I guess because they can't afford it.
[ "Solar evaporation is the oldest and most energy-efficient method of mineral production. At the Great Salt Lake near Ogden, Utah, Compass Minerals draws naturally occurring brine out of the lake into shallow ponds and allows solar evaporation to produce salt, sulfate of potash (SOP) and magnesium chloride. Its SOP plant at the Great Salt Lake is the largest in North America and one of only three SOP brine solar evaporation operations in the world. Annual capacity is 350,000 tons of SOP, 1.5 million tons of salt, and 750,000 tons of magnesium chloride. \n", "Despite huge wind and solar potential, it is too early to say when Morocco could begin exporting renewable electricity to Europe from projects such as the $400 billion Desertec initiative. It is unclear whether the Desertec consortium's planned investment in solar thermal energy across North Africa could go into Morocco or how much power could eventually be exported to Europe. Desertec's plans are likely to need\n", "Some plans exist to build solar farms in the deserts of North Africa to supply power for Europe. The Desertec project, backed by several European energy companies and banks, planned to generate renewable electricity in the Sahara desert and distribute it through a high-voltage grid for export to Europe and local consumption in North-Africa. Ambitions seek to provide continental Europe with up to 15% of its electricity. The TuNur project would supply 2GW of solar generated electricity from Tunisia to the UK.\n", "With a minimum of training in operation and maintenance, solar powered water pumping and purification systems have the potential to help rural Africans fulfill one of their most basic needs for survival. Further field test are in progress by organizations like KARI and the many corporations that manufacture the products needed, and these small-scale applications of solar technology are promising. Combined with sustainable agricultural practices and conservation of natural resources, solar power is a prime candidate to bring the benefits of technology to the parched lands of Africa.\n", "Simple evaporative cooling devices such as evaporative cooling chambers (ECCs) and clay pot coolers, or pot-in-pot refrigerators,  are simple and inexpensive ways to keep vegetables fresh without the use of electricity. Several hot and dry regions throughout the world could potentially benefit from evaporative cooling, including North Africa, the Sahel region of Africa, the Horn of Africa, southern Africa, the Middle East, arid regions of South Asia, and Australia. Benefits of evaporative cooling chambers for many rural communities in these regions include reduced post-harvest loss, less time spent traveling to the market, monetary savings, and increased availability of vegetables for consumption.\n", "Off-grid solar-powered desalination units use solar energy to fill a buffer tank on a hill with seawater. The reverse osmosis process receives its pressurized seawater feed in non-sunlight hours by gravity, resulting in sustainable drinking water production without the need for fossil fuels, an electricity grid or batteries.Nano-tubes are also used for the same function (i.e, Reverse Osmosis).\n", "Solar water disinfection is a low-cost method of purifying water that can often be implemented with locally available materials. Unlike methods that rely on firewood, it has low impact on the environment.\n" ]
how fresh water fish populations spread from one river network to another without going through (presumably deadly) seawater?
I wondered that about fresh water mussels. So I asked. They have a free living form which basically hitches a ride in the gills of fish. This means the same species can be in one river system even though they are mussels. Similar events can occur for other species. Egrets and other fish eating birds will fly from one river system to another. Over geological time species can be transferred. All watersheds border each other so it is not travel over hundreds of miles but perhaps ten miles. Larva can remain in a bird's mouth during such a flight.
[ "A second potential mechanism for accidental movement of infected fish is the international trade in ornamental or aquarium fishes, which includes the global trade of approximately 5000 freshwater and 1450 saltwater fishes. Each year over 1 billion individual fish are shipped among more than 100 nations, creating a serious concern for the spread of megalocytiviruses as well as other important fish pathogens. There is already substantial evidence of this problem: megalocytiviruses which are genetically identical or extremely similar to ISKNV have been isolated from ornamental fishes (gouramis) that were being traded internationally. Furthermore, an Australian outbreak of megalocytivirus among farmed Murray cod (\"Maccullochella peelii\") was linked to imported gouramis in pet shops. In addition, a 2008 study reported 10 aquarium fish species that tested positive for ISKNV in Korea.\n", "The European Commission (2002) concluded “The reduction of wild salmonid abundance is also linked to other factors but there is more and more scientific evidence establishing a direct link between the number of lice-infested wild fish and the presence of cages in the same estuary.” It is reported that wild salmon on the west coast of Canada are being driven to extinction by sea lice from nearby salmon farms. Antibiotics and pesticides are often used to control the diseases and parasites, as well as lasers.\n", "Flooding and pollution have affected fish populations on the river and in Lake St. Lawrence. Northern Pike, Walleye, Muskellunge, Lake Sturgeon and American eel have been affected. The loss of spawning grounds is also believed to have contributed to drops in their populations. Recent efforts have stabilized or increased much of the populations. R.H. Saunders Generating Station has a ladder made in a decommissioned ice sluce designed for juvenile American Eels to head upriver, across the generating station. At long and high, it was the only one in North America and the tallest in the world at the time. In recent years, it has been upgraded and extended in length. OPG maintains a trap and transport program with local commercial fisherman for downstream migration. From 2006 to 2011, approximately four million young eels crossed into the upper St. Lawrence River and Lake Ontario.\n", "Inland waters are populated by small to large fresh water fish, such as the largemouth bass, the American pickerel, the walleye, the \"Acipenser oxyrinchus\", the muskellunge, the Atlantic cod, the Arctic char, the brook trout, the \"Microgadus tomcod\" (tomcod), the Atlantic salmon, the rainbow trout, etc.\n", "The main fish that live in this river is the common trout. It has been confirmed that the 95% of this species, opposite from the salmon, do not reach the sea. Also the 95% of the common trout that live in this river have an average movement of fifty meters. Which means that only a 5% of them travel for distances longer than 500 meters.\n", "Upstream populations have been isolated and decimated by poorly managed grazing and excessive water withdrawals for irrigation, as well as by hybridization, competition, and predation by non-native salmonids. This is important, as although Lahontan cutthroat trout can inhabit either lakes or streams, they are obligatory stream spawners.\n", "Like in many rivers around Europe, there are some introduced salmonid fish species in the Neretva. Of these only grayling (\"Thymallus thymallus\") established stable population so far, while the more harmful rainbow trout (\"Oncorhynchus mykiss\") had lower survival rate and accordingly low population growth and small size. Brook trout (\"Salvelinus fontinalis\") and lake trout (\"Salvelinus namaycush\") have also recently been introduced to almost all of the Neretva basin reservoirs, but had only moderate to low success in establishing stable populations. At least for now populations of these invasive salmonids are rather weak.\n" ]
Do we experience atmosphere tides? Do the molecules in the air get dragged according to the moons gravity causing “deeper” periods of time?
The moon does indeed effect the atmosphere, but orders of magnitude less than it does the ocean. Tidal effects form pressure waves of about 100 microbars, or about 0.01% the atmospheric pressure at sea level. That's only perceptible to scientific instruments, and practically background noise compared to the regular variations in atmospheric pressure that occur due to weather and the effects of the sun. _URL_0_ _URL_1_
[ "Atmospheric tides are global-scale waves excited by regular solar differential heating (thermal tides) or by the gravitational tidal force of the moon (gravitational tides). The atmosphere behaves like a huge waveguide closed at the bottom (the Earth's surface) and open to space at the top. In such a waveguide an infinite number of atmospheric wave modes can be excited. Because the waveguide is imperfect, however, only modes of lowest degree with large horizontal and vertical scales can develop sufficiently well so that they can be filtered out from the meteorological noise. They are solutions of the Laplace equation and are called Hough functions. These can be approximated by a sum of spherical harmonics.\n", "At ground level, atmospheric tides can be detected as regular but small oscillations in surface pressure with periods of 24 and 12 hours. However, at greater heights, the amplitudes of the tides can become very large. In the mesosphere (heights of ~ 50–100 km) atmospheric tides can reach amplitudes of more than 50 m/s and are often the most significant part of the motion of the atmosphere.\n", "Atmospheric tides are also produced through the gravitational effects of the Moon. \"Lunar (gravitational) tides\" are much weaker than \"solar (thermal) tides\" and are generated by the motion of the Earth's oceans (caused by the Moon) and to a lesser extent the effect of the Moon's gravitational attraction on the atmosphere.\n", "i) Atmospheric tides are primarily excited by the Sun's heating of the atmosphere whereas ocean tides are primarily excited by the Moon's gravitational field. This means that most atmospheric tides have periods of oscillation related to the 24-hour length of the solar day whereas ocean tides have longer periods of oscillation related to the lunar day (time between successive lunar transits) of about 24 hours 51 minutes.\n", "Atmospheric tides are negligible at ground level and aviation altitudes, masked by weather's much more important effects. Atmospheric tides are both gravitational and thermal in origin and are the dominant dynamics from about , above which the molecular density becomes too low to support fluid behavior.\n", "BULLET::::1. Atmospheric tides are primarily excited by the Sun's heating of the atmosphere whereas ocean tides are excited by the Moon's gravitational pull and to a lesser extent by the Sun's gravity. This means that most atmospheric tides have periods of oscillation related to the 24-hour length of the solar day whereas ocean tides have periods of oscillation related both to the solar day as well as to the longer lunar day (time between successive lunar transits) of about 24 hours 51 minutes.\n", "Atmospheric tides are global-scale periodic oscillations of the atmosphere. In many ways they are analogous to ocean tides. Atmospheric tides form an important mechanism for transporting energy input into the lower atmosphere from the upper atmosphere, while dominating the dynamics of the mesosphere and lower thermosphere. Therefore, learning about atmospheric tides is essential in understanding the atmosphere as a whole. Modeling and observations of atmospheric tides are needed in order to monitor and predict changes in the Earth's atmosphere.\n" ]
what is the difference between losing weight because i am not eating any food vs losing weight while maintaining a healthy diet and working out to lose more calories?
Losing weight is easy by not eating. But you will feel weak from lack of nutrition and you may still not have a good body composition. Working out and eating well you will feel energised and strong and will end up with a healthy body composition. Body composition is how much fat and muscle you have in your body. You may maintain the same weight your whole life but if you never workout your body composition will change as your muscle atrophy with age. Working out will help maintain your body, your mind and strength so you will still be able to tie your own shoes when you're 80.
[ "Some popular beliefs attached to weight loss have been shown to either have less effect on weight loss than commonly believed or are actively unhealthy. According to Harvard Health, the idea of metabolism being the \"key to weight\" is \"part truth and part myth\" as while metabolism does affect weight loss, external forces such as diet and exercise have an equal effect. They also commented that the idea of changing one's rate of metabolism is under debate. Diet plans in fitness magazines are also often believed to be effective, but may actually be harmful by limiting the daily intake of important calories and nutrients which can be detrimental depending on the person and are even capable of driving individuals away from weight loss.\n", "GEBN's view of weight and metabolic health promoted the idea that weight loss can be achieved by taking more exercise while maintaining the same level of consumption - this view \"crosses a line by advancing a view that falls outside the scientific consensus\", and presents an overly simplistic view of the energy balance equation, with experts noting that \"evidence for eating less as a weight-loss strategy is much, much stronger than the evidence for moving more\".\n", "Weight loss is achieved by adopting a lifestyle in which fewer calories are consumed than are expended. According to the UK National Health Service this is best achieved by monitoring calories eaten and supplementing this with physical exercise.\n", "One of the most important things to take into consideration when either trying to lose or put on weight is output versus input. It is important to know the amount of energy your body is using every day, so that your intake fits the needs of one's personal weight goal. Someone wanting to lose weight would want a smaller energy intake than what they put out. There is increasing research-based evidence that low-fat vegetarian diets consistently lead to healthy weight loss and management, a decrease in diabetic symptoms as well as improved cardiac health.\n", "A study published in \"American Psychologist\" found that short-term dieting involving \"severe restriction of calorie intake\" does not lead to \"sustained improvements in weight and health for the majority of individuals\". Other studies have found that the average individual maintains some weight loss after dieting. Weight loss by dieting, while of benefit to those classified as unhealthy, may slightly increase the mortality rate for individuals who are otherwise healthy.\n", "A commonly asserted \"rule\" for weight gain or loss is based on the assumption that one pound of human fat tissue contains about 3,500 kilocalories (often simply called \"calories\" in the field of nutrition). Thus, eating 500 fewer calories than one needs per day should result in a loss of about a pound per week. Similarly, for every 3500 calories consumed above the amount one needs, a pound will be gained.\n", "Food energy intake must be balanced with activity to maintain a proper body weight. Sedentary individuals and those eating less to lose weight may suffer malnutrition if they eat food supplying empty calories but not enough nutrients. In contrast, people who engage in heavy physical activity need more food energy as fuel, and so can have a larger amount of calorie-rich, essential nutrient-poor foods. Dietitians and other healthcare professionals prevent malnutrition by designing eating programs and recommending dietary modifications according to patient's needs.\n" ]
why are blood stains on fabric so difficult to remove compared to other types of stains?
Mostly because it's a complex substance, containing liquid, suspended solids, cells, fats, oils, and a plethora of other compounds. Most other stains are caused by small portions of these types of compounds. Wine is mostly aqueous (water based), grease is lipid (fat) based, etc. Most solvents are good at cleaning up one type of stain because they designed to pick up one type. Blood, being a mixture of most or all of these types, means many solvents won't work on all of blood's components.
[ "For tissues that are not directly acidic or basic, it can be difficult to use only one stain to reveal the necessary structures of interest. A combination of the three different stains in precise amounts applied in the correct order reveals the details selectively. This is the result of more than just electrostatic interactions of stain with the tissue and the stain not being washed out after each step. Collectively the stains compliment one another.\n", "Gel stains are a late 20th century innovation in stain manufacturing, in that they are high-viscosity liquids and do not 'flow'. This property allows more control during application, particularly when the wood is in a vertical position, which can often cause traditional liquid stains to run, drip, or pool. Gel stains often have limited penetrating ability, as they are thixotropic (a liquid that nevertheless does not 'flow').\n", "Another factor in stain removal is the fact that stains can sometimes comprise two separate staining agents, which require separate forms of removal. A machine oil stain could also contain traces of metal, for example.Also of concern is the color of the material that is stained. Some stain removal agents will not only dissolve the stain, but will dissolve the dye that is used to color the material.\n", "Stains sometimes occur. Some women prefer darker colored fabrics which do not show stains as much as light colored fabrics do. Causes of staining do not include allowing the blood to dry, but using hot water when washing the pad will, as hot water sets protein stains (blood). Often, soaking pads for at least 4-6 hours (or overnight) in cold water with an oxygen bleach can assist in stain removal. Drying cloth pads in sunlight can help to fade stains.\n", "Linen is a very durable, strong fabric, and one of the few that are stronger wet than dry. The fibers do not stretch, and are resistant to damage from abrasion. However, because linen fibers have a very low elasticity, the fabric eventually breaks if it is folded and ironed at the same place repeatedly over time.\n", "Stain is composed of the same three primary ingredients as paint (pigment, solvent (or vehicle), and binder) but is predominantly vehicle, then pigment and/or dye, and lastly a small amount of binder. Much like the dyeing or staining of fabric, wood stain is designed to add color to the substrate (wood and other materials) while leaving some of the substrate still visible. Transparent varnishes or surface films are applied afterwards. In principle, stains do not provide a durable surface coating or film. However, because the binders are from the same class of film-forming binders that are used in paints and varnishes, some build-up of film occurs.\n", "H&E staining does not always provide enough contrast to differentiate all tissues, cellular structures, or the distribution of chemical substances, and in these cases more specific stains and methods are used.\n" ]
How long does it take to boot up a supercomputer?
The only 'supercomputer' I have access to is a network of 128 computers, each with some xeon processors and lots of ram. These can be booted simultaneously, and boot into a minimally configured Linux environment, which is very quick. The master node then needs to start the cluster management software and register all the compute nodes and you're more or less good to go.
[ "This process took a total of 0.337 seconds of CPU time, out of which 0.327 seconds was spent in user space, and the final 0.010 seconds in kernel mode on behalf of the process. Elapsed real time was 1.15 seconds.\n", "To give some perspective to this, using Virginia Tech's System X with a maximum performance of 12.25 Teraflops, it would take approximately formula_8 seconds or about 3 weeks. Or for commodity processors at 2 gigaflops it would take 6,000 machines approximately the same amount of time.\n", "Amdahl's Law reveals a limitation in, for example, the ability of multiple cores to reduce the time it takes for a computer to boot to its operating system and be ready for use. Assuming the boot process was mostly parallel, quadrupling computing power on a system that took one minute to load might reduce the boot time to just over fifteen seconds. But greater and greater parallelization would eventually fail to make bootup go any faster, if any part of the boot process were inherently sequential.\n", "The ability to boot in 30 seconds was a design goal for Windows XP, and Microsoft's developers made efforts to streamline the system as much as possible; The Logical Prefetcher is a significant part of this; it monitors what files are loaded during boot, optimizes the locations of these files on disk so that less time is spent waiting for the hard drive's heads to move and issues large asynchronous I/O requests that can be overlapped with device detection and initialization that occurs during boot. The prefetcher works by tracing frequently accessed paged data which is then used by the \"Task Scheduler\" to create a prefetch-instructions file at %WinDir%\\Prefetch. Once the system boots or an application is started, any data and code specified in the trace that is not already in memory is prefetched from the disk. The previous prefetching results determine which scenario benefited more and what should be prefetched at the next boot or launch. The prefetcher also uses the same algorithms to reduce application startup times. To reduce disk seeking even further, the \"Disk Defragmenter\" is called in at idle time to optimize the layout of these specific files and metadata in a contiguous area. Boot and resume operations can be traced and analyzed using Bootvis.exe.\n", "The default dirty cache delay time was fixed at 2.2 seconds in NetWare 286 versions 2.x. Starting with NetWare 386 3.x, the dirty disk cache delay time and dirty directory cache delay time settings controlled the amount of time the server would cache changed (\"dirty\") data before saving (flushing) the data to a hard drive. The default setting of 3.3 seconds could be decreased to 0.5 seconds but not reduced to zero, while the maximum delay was 10 seconds. The option to increase the cache delay to 10 seconds provided a significant performance boost. Windows 2000 and 2003 server do not allow adjustment to the cache delay time. Instead, they use an algorithm that adjusts cache delay.\n", "Proposed by E. B. Davies, this is a machine that can, in the space of half an hour, create an exact replica of itself that is half its size and capable of twice its replication speed. This replica will in turn create an even faster version of itself with the same specifications, resulting in a supertask that finishes after an hour. If, additionally, the machines create a communication link between parent and child machine that yields successively faster bandwidth and the machines are capable of simple arithmetic, the machines can be used to perform brute-force proofs of unknown conjectures. However, Davies also points out that due to fundamental properties of the real universe such as quantum mechanics, thermal noise and information theory his machine can't actually be built.\n", "In an OS environment on a 360/67 with spooled input and output files, the compiler will recompile itself in about 25 seconds. The compiler is approximately 2700 card images. Thus, when the OS scheduler time is subtracted from the execution time given above, it is seen that the compiler runs at a speed in excess of 100 cards per second (for dense code).\n" ]
Does potential energy "count" as energy?
I somehow miss your sub-question, and I gave an answer. I will keep some of what I wrote there, maybe you will like it. Maybe you didn't realize it, but gravitational energy is also consider/named potential energy. Why? because it has the potential to generate a force (gravitational force in this case) if you put an object in a gravitational field, it will be attracted to the effective force generated from all the gravitational potential in there. Why is it so intriguing?, well it has point out that if you count this potential energy into the evolution of the universe, we could reach a total amount of 0 energy at the beginning. And that is outstanding. Because it means that the Universe was created from nothing. Edit: Read DocSmile answer
[ "Common types of potential energy include the gravitational potential energy of an object that depends on its mass and its distance from the center of mass of another object, the elastic potential energy of an extended spring, and the electric potential energy of an electric charge in an electric field. The unit for energy in the International System of Units (SI) is the joule, which has the symbol J.\n", "The potential energy, \"U\", depends on the position of an object subjected to a conservative force. It is defined as the object's ability to do work and is increased as the object is moved in the opposite direction of the direction of the force. If \"F\" represents the conservative force and \"x\" the position, the potential energy of the force between the two positions \"x\" and \"x\" is defined as the negative integral of \"F\" from \"x\" to \"x\":\n", "Thus, the total energy of the particle—the sum of its kinetic energy and its potential energy \"U\"—is a constant; energy is said to be conserved. To show this, it suffices that the work \"W\" done by the force depends only on initial and final positions, not on the path taken between them.\n", "Potential energy is associated with forces that act on a body in a way that the total work done by these forces on the body depends only on the initial and final positions of the body in space. These forces, that are called \"conservative forces\", can be represented at every point in space by vectors expressed as gradients of a certain scalar function called \"potential\".\n", "In most theoretical physics such as quantum field theory, the energy that a particle has as a result of changes that it itself causes in its environment defines self-energy formula_1, and represents the contribution to the particle's energy, or effective mass, due to interactions between the particle and its system. In electrostatics, the energy required to assemble the charge distribution takes the form of self-energy by bringing in the constituent charges from infinity, where the electric force goes to zero. In a condensed matter context relevant to electrons moving in a material, the self-energy represents the potential felt by the electron due to the surrounding medium's interactions with it. Since electrons repel each other the moving electron polarizes, or causes to displace, the electrons in its vicinity and then changes the potential of the moving electron fields. These and other effects entail self-energy. \n", "This equation shows that the total energy of a particle (bradyon or tachyon) contains a contribution from its rest mass (the \"rest mass–energy\") and a contribution from its motion, the kinetic energy.\n", "This equation shows that the total energy of a particle (bradyon or tachyon) contains a contribution from its rest mass (the \"rest mass–energy\") and a contribution from its motion, the kinetic energy.\n" ]
prism and the nsa scandal
A low level contractor working for the NSA named Snowden leaked a document describing a classified program that collects data from the largest internet communities in the world, including microsoft, facebook, yahoo, google, etc. The internet companies themselves claim to be unaware of this. It's a big deal because the government is throwing a wide net and capturing *all* data rather than targeting who, what or when. The part of the law that deals with this would normally require a warrant (of sorts) before digging into your information. Instead they decided they'd search you first then ask for permission. Especially troubling is that the program searches citizens as well, which is illegal and not sactioned under FISA. The government, under pressure, eventually admitted to the program existing. So it isn't a question or conspiracy, it's very real. Here is a great start / quick reading to explain : _URL_0_
[ "Documents indicate that PRISM is \"the number one source of raw intelligence used for NSA analytic reports\", and it accounts for 91% of the NSA's Internet traffic acquired under FISA section 702 authority.\" The leaked information came to light one day after the revelation that the FISA Court had been ordering a subsidiary of telecommunications company Verizon Communications to turn over to the NSA logs tracking all of its customers' telephone calls.\n", "PRISM is a program from the Special Source Operations (SSO) division of the NSA, which in the tradition of NSA's intelligence alliances, cooperates with as many as 100 trusted U.S. companies since the 1970s. A prior program, the Terrorist Surveillance Program, was implemented in the wake of the September 11 attacks under the George W. Bush Administration but was widely criticized and challenged as illegal, because it did not include warrants obtained from the Foreign Intelligence Surveillance Court. PRISM was authorized by the Foreign Intelligence Surveillance Court.\n", "PRISM is a code name for a program under which the United States National Security Agency (NSA) collects Internet communications from various US Internet companies. The program is also known by the SIGAD . PRISM collects stored Internet communications based on demands made to Internet companies such as Google LLC under Section 702 of the FISA Amendments Act of 2008 to turn over any data that match court-approved search terms. The NSA can use these PRISM requests to target communications that were encrypted when they traveled across the Internet backbone, to focus on stored data that telecommunication filtering systems discarded earlier, and to get data that is easier to handle, among other things.\n", "On June 6, 2013, former NSA contractor Edward Snowden leaked the existence of PRISM, an electronic surveillance program intended to monitor e-mail and phone call activity in the United States to identify possible terrorist threats, to the newspapers \"The Guardian\" and \"The Washington Post\".\n", "In June 2013, Edward Snowden—at the time a Booz Allen employee contracted to projects of the National Security Agency (NSA)—publicly disclosed details of classified mass surveillance and data collection programs, including PRISM. The alleged leaks are said to rank among the most significant breaches in the history of the NSA and led to considerable concern worldwide. Booz Allen condemned Snowden's leak of the existence of PRISM as \"shocking\" and \"a grave violation of the code of conduct and core values of our firm\". The company fired Snowden \"in absentia\" shortly after and stated he had been an employee for less than three months at the time. Market analysts considered the incident \"embarrassing\" but unlikely to cause enduring commercial damage. Booz Allen stated that it would work with authorities and clients to investigate the leak. Charles Riley of \"CNN\"/\"Money\" said that Booz Allen was \"scrambling to distance itself from Snowden\".\n", "PRISM began in 2007 in the wake of the passage of the Protect America Act under the Bush Administration. The program is operated under the supervision of the U.S. Foreign Intelligence Surveillance Court (FISA Court, or FISC) pursuant to the Foreign Intelligence Surveillance Act (FISA). Its existence was leaked six years later by NSA contractor Edward Snowden, who warned that the extent of mass data collection was far greater than the public knew and included what he characterized as \"dangerous\" and \"criminal\" activities. The disclosures were published by \"The Guardian\" and \"The Washington Post\" on June 6, 2013. Subsequent documents have demonstrated a financial arrangement between the NSA's Special Source Operations division (SSO) and PRISM partners in the millions of dollars.\n", "Edward Snowden's leaking of the existence of PRISM in 2013 caused the NSA to institute a \"two-man rule\", where two system administrators are required to be present when one accesses certain sensitive information. Snowden claims he suggested such a rule in 2009.\n" ]
why does rubbing alcohol, vinegar, etc have expiry dates? how can they go bad?
The FDA requires expiry dates on nearly everything aside from alcohol and cosmetics. The manufacturers have to put something, so they generally choose a date a few years out, that they are willing to guarantee the product will meet or exceed that date. _URL_0_
[ "Currently there is no practical way of halting or reversing the course of degradation. Many film collectors use camphor tablets but it is not known what the long term effects on the film would be. While there has been significant research regarding various methods of slowing degradation, such as storage in molecular sieves, temperature and moisture are the two key factors affecting the rate of deterioration. According to the Image Permanence Institute, fresh acetate film stored at a temperature of 70 °F (21 °C) and 40% relative humidity will last approximately 50 years before the onset of vinegar syndrome. Reducing the temperature by 15° while maintaining the same level of humidity brings a dramatic improvement: at a temperature of 55 °F (13 °C) and 40% relative humidity, the estimated time until onset of vinegar syndrome is 150 years. A combination of low temperature and low relative humidity represents the optimum storage condition for cellulose acetate base films, with the caveat that relative humidity should not be lowered below 20%, or the film will dry out too much and become brittle.\n", "Alcohol rub sanitizers kill most bacteria, and fungi, and stop some viruses. Alcohol rub sanitizers containing at least 70% alcohol (mainly ethyl alcohol) kill 99.9% of the bacteria on hands 30 seconds after application and 99.99% to 99.999% in one minute.\n", "A dilute alcohol solution inoculated with \"Acetobacter\" and kept in a warm, airy place will become vinegar over the course of a few months. Industrial vinegar-making methods accelerate this process by improving the supply of oxygen to the bacteria.\n", "The spent hops are removed from the bitter liquor by decanting or 'casting' it into the hop back, a vessel on the ground floor beneath the coppers. The spent hops settle out and the liquor is strained through them.\n", "Reggio Emilia designates the different ages of their balsamic vinegar (\"Aceto Balsamico Tradizionale di Reggio Emilia\") by label colour. A red label means the vinegar has been aged for at least 12 years, a silver label that the vinegar has aged for at least 18 years, and a gold label designates that the vinegar has aged for 25 years or more.\n", "Product labels for rubbing alcohol include a number of warnings about the chemical, including the flammability hazards and its intended use only as a topical antiseptic and not for internal wounds or consumption. It should be used in a well-ventilated area due to inhalation hazards. Poisoning can occur from ingestion, inhalation, absorption, or consumption of rubbing alcohol.\n", "Alcohol is added to mouthwash not to destroy bacteria but to act as a carrier agent for essential active ingredients such as menthol, eucalyptol and thymol which help to penetrate plaque. Sometimes a significant amount of alcohol (up to 27% vol) is added, as a carrier for the flavor, to provide \"bite\". Because of the alcohol content, it is possible to fail a breathalyzer test after rinsing although breath alcohol levels return to normal after 10 minutes. In addition, alcohol is a drying agent, which encourages bacterial activity in the mouth, releasing more malodorous volatile sulfur compounds. Therefore, alcohol-containing mouthwash may temporarily worsen halitosis in those who already have it, or indeed be the sole cause of halitosis in other individuals.\n" ]
In the 60s, was japan's economy expected to overtake the usa's by now?
One minor note: Realistically, China will probably become the world's largest economy much sooner than 2050. The OECD thinks it could be [within the next 4-5 years](_URL_0_), although that's almost certainly jumping the gun. The U.S. National Intelligence Council [thinks it'll be around 2030](_URL_1_) if present trends hold. They're by no means guaranteed to -- there are a lot of problems in China's economy that are currently being hidden by their growth rate, and the U.S. is likely entering a natural resources boom -- but either way, 2050 is probably unrealistic. All other things being equal, China and India *should* be the largest economies on the planet if for no reason other than the size of their internal markets. I feel like India gets lost in this debate a lot. And all of this might turn out to be yet another prediction that never comes true. Macroeconomic tea-reading has never been humanity's strong suit. Echoing the others here, [predictions about Japan's overtaking the U.S.](_URL_2_) were more a feature of the late 1970s and 1980s. The Harvard professor Ezra Vogel's [*Japan as Number One: Lessons for America*](_URL_3_) was published in 1979, for example, and is a somewhat entertaining read in hindsight. Not because Vogel got everything wrong -- a lot of his analysis is spot-on -- but because it's a decent warning against academic overconfidence.
[ "The high economic growth and political tranquillity of the mid-to-late 1960s were tempered by the quadrupling of oil prices by the Organization of the Petroleum Exporting Countries (OPEC) in 1973. Almost completely dependent on imports for petroleum, Japan experienced its first recession since World War II.\n", "After 1950 Japan's economy recovered from the war damage and began to boom, with the fastest growth rates in the world. Given a boost by the Korean War, in which it acted as a major supplier to the UN force, Japan's economy embarked on a prolonged period of extremely rapid growth, led by the manufacturing sectors. Japan emerged as a significant power in many economic spheres, including steel working, car manufacturing and the manufacturing of electronics. Japan rapidly caught up with the West in foreign trade, GNP, and general quality of life. The high economic growth and political tranquility of the mid to late 1960s were slowed by the quadrupling of oil prices in 1973. Almost completely dependent on imports for petroleum, Japan experienced its first recession since World War II. Another serious problem was Japan's growing trade surplus, which reached record heights. The United States pressured Japan to remedy the imbalance, demanding that Tokyo raise the value of the yen and open its markets further to facilitate more imports from the United States.\n", "In the 1930s, the Japanese economy suffered less from the Great Depression than most industrialized nations, its GDP expanding at the rapid rate of 5% per year. Manufacturing and mining came to account for more than 30% of GDP, more than twice the value for the agricultural sector. Most industrial growth, however, was geared toward expanding the nation's military power.\n", "The Great Depression, just as in many other countries, hindered Japan's economic growth. The Japanese Empire's main problem lay in that rapid industrial expansion had turned the country into a major manufacturing and industrial power that required raw materials; however, these had to be obtained from overseas, as there was a critical lack of natural resources on the home islands.\n", "The Japanese growth in the postwar period was often called a \"miracle\". It was led by manufacturing; starting with textiles and clothing and moving to high-technology, especially automobiles, electronics and computers. The economy experienced a major slowdown starting in the 1990s following three decades of unprecedented growth, but Japan still remains a major global economic power.\n", "The Japanese financial recovery continued even after SCAP departed and the economic boom propelled by the Korean War abated. The Japanese economy survived from the deep recession caused by a loss of the U.S. payments for military procurement and continued to make gains. By the late 1960s, Japan had risen from the ashes of World War II to achieve an astoundingly rapid and complete economic recovery. According to Knox College Professor Mikiso Hane, the period leading up to the late 1960s saw \"the greatest years of prosperity Japan had seen since the Sun Goddess shut herself up behind a stone door to protest her brother Susano-o's misbehavior.\" The Japanese government contributed to the post-war Japanese economic miracle by stimulating private sector growth, first by instituting regulations and protectionism that effectively managed economic crises and later by concentrating on trade expansion.\n", "Following Japanese defeat, economic collapse after the war, and US military occupation, Japan's economy recovered in the 1950s with the post-war economic miracle in which rapid growth propelled the country to become the world's second largest economy by the 1980s.\n" ]
if one were to strike a billiards ball with the cue ball, and we disregard the deceleration as the target ball hits each bumper as well as other balls on the table, would the ball eventually, but always sink into a pocket regardless of where on the table it went?
No, it won't always sink. It's easy to think of a situation in which it won't happen. If you hit the ball parallel to any side of the table, it would bounce back and forth forever, never entering any of the pockets (unless it was right in front of the pocket already.
[ "BULLET::::- Knocking any ball off the table; opponent receives ball-in-hand plus 2 points (the ball is spotted in its starting position, or as close to this position as possible, unless it was the now-incoming opponent's cue ball, which as noted is in-hand).\n", "Making the ball: You score a point by making the object ball in one of the two corner pockets on your opponent's (i.e., the opposite) side of the table or by banking the ball into a side pocket or back into one of your own pockets. If, upon striking the object ball, it goes either directly into a side pocket or the pockets on your own side of the table then you lose the point (note: toilet paper to block the side pockets is an acceptable variation). If the cue ball goes into ANY pocket on your throw, you lose the point—with one exception, the \"shuck\", explained below.\n", "BULLET::::- Knocking over pins with the object ball without hitting the opponent's cue ball first, or with one's own cue ball, does not earn the shooter any points, and in the latter case is a foul that awards points to the opponent.\n", "BULLET::::- Hitting the pins directly with the shooter's cue ball before any contact with the opponent's cue ball; opponent receives ball-in-hand plus 2 points (the erstwhile value of the knocked-over pins is not calculated at all).\n", "When a ball with (sidespin) on it hits an object ball with a degree of fullness, the object ball will be \"thrown\" in the opposite direction of the side of the cue ball the was applied. Thus, a cue ball with left hand on it will \"throw\" a hit object ball to the right. \n", "BULLET::::- ( in snooker terms) – striking one's cue ball so that it hits another ball and then enters a pocket: 3 points if the red ball was hit first; 2 points if the other cue ball was hit first; 2 points if the red and the other cue ball are hit simultaneously.\n", "Scoring: Once the object ball is moving it is like a \"time-bomb\" in the sense that when it stops moving the location of the cue ball on the table will (may) determine the winner of the point, i.e. if the cue ball is on your side of table when the object ball stops moving, you lose the point. The only variation to this is the \"Gentleman's Rule\" (see below).\n" ]
germany just made all universities free, how do other countries do this?
Economist here. There is no such thing as free. They are just making college payments compulsory for everyone regardless of whether they use it or not. The system used by the US was fine until universities realized the government would subsidize them raising tuition to ridiculous rates by lending to students.
[ "In Germany, most institutions of higher education are subsidised by German states and are therefore also referred to as \"staatliche Hochschulen.\" (public universities) In most German states, admission to public universities is still cheap, about two hundred Euro per semester. In 2005, many states introduced additional fees of 500 Euro per semester to achieve a better teaching-quality; however, all of these states (except Lower Saxony, which will follow in 2014/15) have abolished tuition fees as of autumn 2013. Nevertheless, additional fees for guest or graduate students are charged by many universities.\n", "In Germany, universities are institutions of higher education which have the power to confer bachelor, master and PhD degrees. They are explicitly recognised as such by law and cannot be founded without government approval. The term Universitaet (i.e. the German term for university) is protected by law and any use without official approval is a criminal offense. Most of them are public institutions, though a few private universities exist. Such universities are always research universities. Apart from these universities, Germany has other institutions of higher education (Hochschule, Fachhochschule). Fachhochschule means a higher education institution which is similar to the former polytechnics in the British education system, the English term used for these German institutions is usually 'university of applied sciences'. They can confer master degrees but no PhDs. They are similar to the model of teaching universities with less research and the research undertaken being highly practical. Hochschule can refer to various kinds of institutions, often specialised in a certain field (e.g. music, fine arts, business). They might or might not have the power to award PhD degrees, depending on the respective government legislation. If they award PhD degrees, their rank is considered equivalent to that of universities proper (Universitaet), if not, their rank is equivalent to universities of applied sciences.\n", "In the German Education system almost all universities and most universities of applied sciences are funded by the state and do not charge tuition fees. In exceptional cases universities may offer courses for professionals (e.g. executive MBA programs), which may require tuition payment. Some local governments have recently decided that students from non-EU countries can be charged, although ERASMUS students, students from developing countries and other special groups are exempt. In addition, some private institutions of higher education run on a tuition-based model. \n", "Most of the German universities are public institutions, charging fees of only around €60-200 per semester for each student, usually to cover expenses associated with the university cafeterias and (usually mandatory) public transport tickets. Thus, academic education is open to most citizens and studying is very common in Germany. The dual education system combines both practical and theoretical education but does not lead to academic degrees. It is more popular in Germany than anywhere else in the world and is a role model for other countries.\n", "Most of the German universities are public institutions, and students traditionally study without fee payment. The general requirement for university is the \"Abitur\". However, there are a number of exceptions, depending on the state, the college and the subject. Tuition free academic education is open to international students and is increasingly common.\n", "Germany has both private and public universities with the majority being public universities, which is part of the reason their graduates do not have as much debt. For undergraduate studies, public universities are free but have an enrollment fee of no more than €250 per year, which is roughly US$305. Their private universities cost an average of €10,000 a semester, which is about US$12,000. Private universities account for 7.1% enrollment with the rest attending the public universities. The private universities have a smaller teacher to student ratio and tend to offer more specialized programs, which is why Germany is experiencing a boom in private universities enrollment in recent years for majors like law and medicine. However, most students still prefer public universities due to the drastic difference in tuition cost. The only expense students take out loans for in public universities is the living cost, which ranges from €3600 to €8,200 a year depending on the university location. However, the repayment of this loan is interest free and no borrower pays more than €10,000 regardless of the borrowed amount. The average debt at graduation is €5,600 which is US$6,680. The chance to gain a bachelor's degree through well respected universities at a reasonable price without interest packed loans attracts many foreign students, as seen through increased enrollment of students from all around the world.\n", "As Germany is a member of the European Union, university students have the opportunity to participate in the Erasmus Programme. The university also has exchange programs and partnerships with reputable universities outside Europe such as University of Technology, Sydney in Australia, Tsinghua University, Peking University and Fudan University in China, Tokyo University in Japan and the University of California, Berkeley, in the United States.\n" ]
why apple released ios 6 maps when it obviously wasn't ready for release.
Because google refused to license turn by turn, which was a HUGE feature. Apple had been trying to get turn by turn built in for years but they could never come to an agreement. Also, google changed it's pricing from wholesale licensing to pay-per-use, which would make the maps more expensive. On top of that, google started adding ads to their search results. Apple was also getting nervous that google was collecting user info, so they made their own. Basically they had 2 options: 1. Go another year without turn by turn, potentially losing users to android. Pay more for maps in the meantime as google advertised to their users. 2. Buy a bunch of mapping services, develop their own maps using third party data like tomtom or osm. Wait for the kinks to work themselves out through user reports. They chose option 2. It's really not that hard to grasp.
[ "In iOS 6, Apple replaced Google Maps with its own Apple Maps as the default mapping service for the operating system, and immediately faced criticism for inaccurate or incomplete data, including a museum in a river, missing towns, satellite images obscured by clouds, missing local places, and more.\n", "In June 2016, Eddy Cue said in an interview with \"Fast Company\" that Apple \"had completely underestimated the product, the complexity of it.\" He also said the problems with Apple Maps led to \"significant changes to all of our development processes.\" After the launch of Maps, Apple started offering public betas of new versions of iOS and OS X. Furthermore, Cue commented that before Maps was launched Apple's executive team long discussed if Apple should have its own mapping service. One month later, Tim Cook looked back to the launch of Apple Maps in an interview with \"The Washington Post\" and said \"Maps was a mistake.\" He added that the company admitted its mistake and that Maps is something the company is now proud of because of the improvements.\n", "Before Apple Maps launched as the default mapping application in iOS, Google Maps held this position since the first generation iPhone release in 2007. In late 2009, tensions between Google and Apple started to grow when the Android version of Google Maps featured turn-by-turn navigation, a feature which the iOS version lacked. At the time, Apple argued that Google collected too much user data. When Apple made iOS 6 available, Google Maps could only be accessed by iOS 6 users via the web. Although Google did not immediately launch a mapping application of its own, shortly after the announcement of Apple Maps, Google did add an equivalent of Apple Maps' Flyover feature to its virtual globe application Google Earth. Three months later, in December 2012, Google Maps was released in the App Store. This version of Google Maps, unlike the previous version, featured turn-by-turn navigation. Shortly after it was launched, Google Maps was the most popular free application in the App Store.\n", "On June 11, 2012, during the Apple Worldwide Developers Conference (WWDC), Apple announced the initial release of Apple Maps and revealed that the application would replace Google Maps as the default web mapping service in iOS 6 and beyond. Apple also announced that the application would include turn-by-turn navigation, 3D maps, Flyovers, and the virtual assistant Siri. Furthermore, Apple stated that iPhone users would be able to navigate Apple Maps while in the locked screen. The mapping service was released on September 19, 2012. Following the launch, Apple Maps was heavily criticized, which resulted in a public apology by Apple CEO Tim Cook in late September and the departure of two key employees of Apple. (See also §Early inaccuracy)\n", "Reviewers and commentators were critical of the new Maps app that replaced Google Maps in iOS 6. It had been reported to contain errors such as misplacement of landmark tags, directing users to incorrect locations and poor satellite images. Nine days after Maps' release, Apple issued a statement apologizing for the frustration it had caused customers and recommending that they try alternate mapping services.\n", "iOS 6 features several new and/or updated apps, which includes Apple Maps and Passbook. Apple's built-in Maps app, which replaced the former Maps app powered by Google Maps, had been universally derided and lacked many features present in competing maps apps. It uses Apple's new vector-based engine that eliminates lag, making for smoother zooming. New to Maps is turn-by-turn navigation spoken directions, 3D views in some major cities and real-time traffic. iOS 6 is able to retrieve documents such as boarding passes, admission tickets, coupons and loyalty cards through its new Passbook app. An iOS device with Passbook can be scanned under a reader to process a mobile payment at locations that have compatible hardware. The app has context-aware features such as notifications for relevant coupons when in the immediate vicinity of a given store.\n", "During WWDC in June 2013, Apple announced the new version of Apple Maps in iOS 7. This new version had a new look and icon. A number of new functions were also implemented, including full-screen mode, night mode, real-time traffic information, navigation for pedestrians, and the Frequent Locations feature. The latter feature, which can be switched on and off, was introduced to record the most frequently visited destinations by users in order to improve Apple Maps. In addition, new satellite imagery was added once again. On September 18, 2013, Apple released iOS 7. At that time, the new iPhone 5S included a new motion coprocessor, the M7, which can identify whether a user is walking or driving in order to adjust the navigation mode.\n" ]
How is light that is as old as the universe travelling to us from 13 billion light years away?
The universe was opaque until ~300,000 years after its formation. So the light from the CMB is around that age. Now, imagine an infinitely long ruler. Pick any marking on that ruler and place yourself there. When the universe became transparent, a photon from 10cm from you gets sent in your direction, but due to the universe expanding, the space between markings keep increasing. You and the light source are both still at the original markings, but the expansion has caused that 10cm to become billions of light years. When the photon eventually reaches you, it must appear to originate from that marking which is now far away. P.S. There is no center of the universe.
[ "The light took much longer than 4 billion years to reach us though it was emitted from only 4 billion light years away, and, in fact, the light emitted towards the Earth was actually moving \"away\" from the Earth when it was first emitted, in the sense that the metric distance to the Earth increased with cosmological time for the first few billion years of its travel time, and also indicating that the expansion of space between the Earth and the quasar at the early time was faster than the speed of light. None of this surprising behavior originates from a special property of metric expansion, but simply from local principles of special relativity integrated over a curved surface.\n", "The light reaching Earth from z8_GND_5296 shows its position over 13 billion years ago, having traveled a distance of more than 13 billion light-years. Due to the expansion of the universe, this position is now at about (comoving distance) from Earth.\n", "Receiving light and other signals from distant astronomical sources can even take much longer. For example, it has taken 13 billion (13) years for light to travel to Earth from the faraway galaxies viewed in the Hubble Ultra Deep Field images. Those photographs, taken today, capture images of the galaxies as they appeared 13 billion years ago, when the universe was less than a billion years old. The fact that more distant objects appear to be younger, due to the finite speed of light, allows astronomers to infer the evolution of stars, of galaxies, and of the universe itself.\n", "Since one might not travel faster than light, one might conclude that a human can never travel further from the Earth than 40 light-years if the traveler is active between the age of 20 and 60. A traveler would then never be able to reach more than the very few star systems which exist within the limit of 20–40 light-years from the Earth. This is a mistaken conclusion: because of time dilation, the traveler can travel thousands of light-years during their 40 active years. If the spaceship accelerates at a constant 1 g (in its own changing frame of reference), it will, after 354 days, reach speeds a little under the speed of light (for an observer on Earth), and time dilation will increase their lifespan to thousands of Earth years, seen from the reference system of the Solar System, but the traveler's subjective lifespan will not thereby change. If the traveler returns to the Earth, they will land thousands of years into the Earth's future. Their speed will not be seen as higher than the speed of light by observers on Earth, and the traveler will not measure their speed as being higher than the speed of light, but will see a length contraction of the universe in their direction of travel. And as the traveler turns around to return, the Earth will seem to experience much more time than the traveler does. So, while their (ordinary) coordinate speed cannot exceed \"c\", their proper speed (distance as seen by Earth divided by their proper time) can be much greater than \"c\". This is seen in statistical studies of muons traveling much further than \"c\" times their half-life (at rest), if traveling close to \"c\".\n", "Older studies generally derived smaller distances such as 345, 370, or 430 lights years. The original parallax calculated from Hipparcos measurements was 9.43 mas, indicating a distance of 346 light years.\n", "Astronomical distances are sometimes expressed in light-years, especially in popular science publications and media. A light-year is the distance light travels in one year, around 9461 billion kilometres, 5879 billion miles, or 0.3066 parsecs. In round figures, a light year is nearly 10 trillion kilometres or nearly 6 trillion miles. Proxima Centauri, the closest star to Earth after the Sun, is around 4.2 light-years away.\n", "This distance is the time (in years) that it took light to reach the observer from the object multiplied by the speed of light. For instance, the radius of the observable universe in this distance measure becomes the age of the universe multiplied by the speed of light (1 light year/year) i.e. 13.8 billion light years. Also see misconceptions about the size of the visible universe.\n" ]
why is the minimum age for ‘adult’ medicines usually 12? not 18 or 21?
Because a lot of pharmaceuticals are dosed by patient weight. At 12 years old you are nearly full size.
[ "The most frequently reported drugs that have been associated with the development of AGEP include penicillin, aminopenicillins, macrolides, quinolones, sulfonamides, hydroxychloroquine, terbinafine, and diltiazem. A more complete list of drugs sorted by their intended actions are:\n", "The initiation of drug and alcohol use is most likely to occur during adolescence, and some experimentation with substances by older adolescents is common. For example, results from 2010 Monitoring the Future survey, a nationwide study on rates of substance use in the United States, show that 48.2% of 12th graders report having used an illicit drug at some point in their lives. In the 30 days prior to the survey, 41.2% of 12th graders had consumed alcohol and 19.2% of 12th graders had smoked tobacco cigarettes. In 2009 in the United States about 21% of high school students have taken prescription drugs without a prescription. And earlier in 2002, the World Health Organization estimated that around 140 million people were alcohol dependent and another 400 million with alcohol-related problems.\n", "In the United States, nicotine products and Nicotine Replacement Therapy products like Nicotrol are only available to persons 18 and above; proof of age is required; not for sale in vending machine or from any source where proof of age cannot be verified. In some states, these products are only available to persons over the age of 21.\n", "In April 2011, the UK National Research Ethics Service approved prescribing monthly injection of puberty-blocking drugs to youngsters from 12 years old, in order to enable them to get older before deciding on formal sex change. The Tavistock and Portman NHS Foundation Trust (T&P) in North London has treated such children. Clinic director Dr. Polly Carmichael said, \"Certainly, of the children between 12 and 14, there's a number who are keen to take part. I know what's been very hard for their families is knowing that there's something available but it's not available here.\" The clinic received 127 referrals for gender dysphoria in 2010.\n", "BULLET::::- Old age: factors relating to how human physiology changes with age may affect the interaction of drugs. For example, liver metabolism, kidney function, nerve transmission or the functioning of bone marrow all decrease with age. In addition, in old age there is a sensory decrease that increases the chances of errors being made in the administration of drugs.\n", "Children under the age of 18 are regarded as underage children. Underage children up to the age of 12 may be included in the Passport of their parents or acquire their own Passport provided both parents consent.\n", "Among US adults older than 55, 4% are taking medication and or supplements that put them at risk of a major drug interaction. Potential drug-drug interactions have increased over time and are more common in the low educated elderly even after controlling for age, sex, place of residence, and comorbidity.\n" ]
How long after sperm enters the egg does it take to create the DNA for the baby?
The term you're looking for is the "zygote"; this is the first cell that properly has all 46 chromosomes. In general, this takes about 12 hours to form after the sperm cell has entered the ovum. As to whether it's the "start of a life", I guess you'd have to define what life is. Both the sperm and oocyte prior to fertilization are "alive", and the zygote that's made is also "alive". Perhaps you mean "human life"? Well then you'd need to define a human.
[ "In contrast to oogenesis, the production of sperm cells is a lifelong process. Each year after puberty, spermatogonia (precursors of the spermatozoa) divide meiotically about 23 times. By the age of 40, the spermatogonia will have undergone about 660 such divisions, compared to 200 at age 20. Copying errors might sometimes happen during the DNA replication preceding these cell divisions, which may lead to new (\"de novo\") mutations in the sperm DNA.\n", "During conception, the father's sperm cell and the mother's egg cell, each containing half the amount of DNA found in other body cells, meet and fuse to form a fertilized egg, called a zygote. The zygote contains a complete set of DNA molecules, a unique combination of DNA from both parents. This zygote divides and multiplies into an embryo and later, a full human being.\n", "The sperm and the egg cell, which has been released from one of the female's two ovaries, unite in one of the two fallopian tubes. The fertilized egg, known as a zygote, then moves toward the uterus, a journey that can take up to a week to complete. Cell division begins approximately 24 to 36 hours after the female and male cells unite. Cell division continues at a rapid rate and the cells then develop into what is known as a blastocyst. The blastocyst arrives at the uterus and attaches to the uterine wall, a process known as implantation.\n", "The sperm is stored in the female's sperm storage tubules for a period varying from a week to more than 100 days, depending on the species. Then, eggs will be fertilized individually as they leave the ovaries, before the shell is calcified in the oviduct. After the egg is laid by the female, the embryo continues to develop in the egg outside the female body.\n", "4 hours after fusion of sperm and ovum, DNA synthesis begins. Male and female pronuclei move to the centre of the egg and membranes break down. Male protamines are replaced with histones and the male DNA is demethylated. Chromosomes then orientate on the metaphase spindle for mitosis. This combination of the two genomes is called syngamy.\n", "Spermatogenesis continues after birth. In the third to fifth months of life, some of the fetal spermatogonia residing along the basement membrane become type A spermatogonia. More gradually, other fetal spermatogonia become type B spermatogonia and primary spermatocytes by the fifth year after birth. Spermatogenesis arrests at this stage until puberty.\n", "The gestation period lasts for about eleven months, or about 340 days (normal average range 320–370 days). During the early days of pregnancy, the conceptus is mobile, moving about in the uterus until about day 16 when \"fixation\" occurs. Shortly after fixation, the embryo proper (so called up to about 35 days) will become visible on trans-rectal ultrasound (about day 21) and a heartbeat should be visible by about day 23. After the formation of the endometrial cups and early placentation is initiated (35–40 days of gestation) the terminology changes, and the embryo is referred to as a fetus. True implantation – invasion into the endometrium of any sort – does not occur until about day 35 of pregnancy with the formation of the endometrial cups, and true placentation (formation of the placenta) is not initiated until about day 40-45 and not completed until about 140 days of pregnancy. The fetus's sex can be determined by day 70 of the gestation using ultrasound. Halfway through gestation the fetus is the size of between a rabbit and a beagle. The most dramatic fetal development occurs in the last 3 months of pregnancy when 60% of fetal growth occurs.\n" ]
What connection is there between the esophageal sphincter and your ears?
Are you referring to the upper esophageal sphincter or the lower?
[ "The external ear openings are covered with fur and do not have a pinnae. The nostrils are small vertical slits right below the shield-like rostrum. Although the brain has been regarded as very primitive and represents the \"lowliest marsupial brain\", the olfactory bulbs and the rubercula olfactoria are very well developed. This seems to suggest that the olfactory sense plays an important role in the marsupial moles' life, as it would be expected for a creature living in an environment lacking visual stimuli. The middle ear seems to be adapted for the reception of low-frequency sounds.\n", "The superior conchae completely cover and protect the nerve axons piercing through the cribriform plate (a porous bone plate that separates the nose from the brain) into the nose. Some areas of the middle conchae are also innervated by the olfactory bulb. All three pairs of conchae are innervated by pain and temperature receptors, via the trigeminal nerve (or, the fifth cranial nerve). Research has shown that there is a strong connection between these nerve endings and activation of the olfactory receptors, but science has yet to fully explain this interaction.\n", "Points located on the ear lobe are related to the head and facial region, those on the scapha are related to the upper limbs, while those on the antihelix and anihelix crura to the trunk and lower limbs, and those in the concha are related to the internal organs.\n", "In most vertebrates, the olfactory bulb is the most rostral (forward) part of the brain, as seen in rats. In humans, however, the olfactory bulb is on the inferior (bottom) side of the brain. The olfactory bulb is supported and protected by the cribriform plate of the ethmoid bone, which in mammals separates it from the olfactory epithelium, and which is perforated by olfactory nerve axons. The bulb is divided into two distinct structures: the main olfactory bulb and the accessory olfactory bulb.\n", "The cochlea is the tri-chambered auditory detection portion of the ear, consisting of the scala media, the scala tympani, and the scala vestibuli. Regarding mammals, placental and marsupial cochleae have similar cochlear responses to auditory stimulation as well as DC resting potentials. This leads to the investigation of the relationship between these therian mammals and researching their ancestral species to trace the origin of the cochlea.\n", "There are six receptor organs located in the inner ear: cochlea, utricle, saccule, and the lateral, anterior, and posterior semicircular canals. The cochlea is a sensory organ with the primary purpose to aid in hearing. The otolith organs (utricle and saccule) are sensors for detecting linear acceleration in their respective planes (utrical=horizontal plane (forward/backward; up/down); saccule=sagital plane (up/down)), and the three semicircular canals (anterior/superior, posterior, and horizontal) detect head rotation or angular acceleration in their respective planes of orientation (anterior/superior=pitch (nodding head), posterior=roll (moving head from one shoulder to other), and horizontal=yaw (shaking head left to right). \n", "Prominent ear, otapostasis or bat ear is an abnormally protruding human ear. It may be unilateral or bilateral. The concha is large with poorly developed antihelix and scapha. It is the result of malformation of cartilage during primitive ear development in intrauterine life. The deformity can be corrected anytime after 6 years. The surgery is preferably done at the earliest in order to avoid psychological distress. Correction by otoplasty involves changing the shape of the ear cartilage so that the ear is brought closer to the side of the head. The skin is not removed, but the shape of the cartilage is altered. The surgery does not affect hearing. It is done for cosmetic purposes only. The complications of the surgery, though rare, are keloid formation, hematoma formation, infection and asymmetry between the ears.\n" ]
farting and its relation to poop.
> what purpose does farting serve? Farting serves the purpose of releasing excess gas in your digestive system. These gases are generally produced by (beneficial to you) bacteria that live within your digestive system > Why do they smell identical to the shit that I would imminently blast out? A human perceives smell by directly having relevant particles enter the nose. When you smell poop, you're detecting poop-particles entering your nose. When you smell farts, you're detecting those same poop-particles entering your nose.
[ "Originally, the various Mandarin Chinese words for \"excrement\" were less commonly used as expletives, but that is changing. Perhaps because farting results in something that is useless even for fertilizer: \"fàng pì\" (; lit. \"to fart\") is an expletive in Mandarin. The word \"pì\" (; lit. \"fart\") or the phrase is commonly used as an expletive in Mandarin (i.e. \"bullshit!\").\n", "The Gas We Pass: The Story of Farts (おなら \"Onara\") is a children's book written by Shinta Chō (). It was first published in Japan in 1978; the first American edition was in 1994. The book tells children about flatulence (also known as farting), and that it is completely natural to do so. \n", "Flatulence or farting works much like burping, but helps the body pass gas through the anus, not the mouth. Bacteria present in the intestinal tract cause gas to be expelled from the anus. They produce the gas as food is digested and moved from the small intestine. This gas builds up and causes swelling or bloating in the abdominal area before it is released.\n", "Using the term \"shit\" (or other locally used crude words) – rather than feces or excreta – during campaigns and triggering events is a deliberate aspect of the community-led total sanitation approach which aims to stop open defecation, a massive public health problem in developing countries.\n", "The sourcing of a fart involves a ritual of assignment that sometimes takes the form of a rhyming game. These are frequently used to discourage others from mentioning the fart or to turn the embarrassment of farting into a pleasurable subject matter. The trick is to pin the blame on someone else, often by means of deception, or using a back and forth rhyming game that includes phrases such as the following:\n", "In the word's literal sense, it has a rather small range of common usages. An unspecified or collective occurrence of feces is generally \"shit\" or \"some shit\"; a single deposit of feces is sometimes \"a shit\" or \"a piece of shit\"; and to defecate is \"to shit\" or \"to take a shit\". While it is common to speak of shit as existing in \"a pile\", \"a load\", \"a hunk\", and other quantities and configurations, such expressions flourish most strongly in the figurative. For practical purposes, when actual defecation and excreta are spoken of, it is either through creative euphemism or with a vague and fairly rigid literalism.\n", "According to \"The Alphabet of Manliness\", the assigning of blame for farting is part of a ritual of behaviour. This may involve deception and a back and forth rhyming game, for example, \"He who smelt it, dealt it\" and \"He who denied it, supplied it\". Derived terms include \"fanny fart\" (queef), \"brain fart\" (slang for a special kind of abnormal brain activity which results in human error while performing a repetitive task, or more generally denoting a degree of mental laxity or any task-related forgetfulness, such as forgetting how to hold a fork) and \"old fart\".\n" ]
Why did none of the southern states have ballots for Lincoln in the 1860 election?
Something to keep in mind is that printed, government-supplied ballots happened relatively late in our political process. In 1860, one would vote by writing out a ballot for a candidate or slate of candidates, or alternatively by using a ballot that was printed in a newspaper or similar publication or handed out by a candidate's supporters at the polls. So Lincoln not appearing on ballots in the South is fairly normal; he was so unpopular there that few people would have dared to write him in or publish a ballot for him.
[ "Among the slave states, the three states with the highest voter turnouts voted the most one-sided. Texas, with five percent of the total wartime South's population, voted 75 percent Breckinridge. Kentucky and Missouri, with one-fourth the total population, voted 73 percent pro-union Bell, Douglas and Lincoln. In comparison, the six states of the Deep South making up one-fourth the Confederate voting population, split 57 percent Breckinridge versus 43 percent for the two pro-union candidates. The four states that were admitted to the Confederacy after Fort Sumter held almost half its population, and voted a narrow combined majority of 53 percent for the pro-union candidates.\n", "In the United States presidential election of 1860 voting was sharply divided, with the south voting for the Southern Democrats and the north for Abraham Lincoln's Republican Party. Lincoln was anti-slavery, did not acknowledge the right to secession, and would not yield federal property in Southern states. Southern secessionists believed Lincoln's election meant long-term doom for their slavery-based agrarian economy and social system.\n", "Despite his early fears of defeat, Lincoln won strong majorities in the popular and electoral vote, partly as a result of the recent Union victory at the Battle of Atlanta. As the Civil War was still raging, no electoral votes were counted from any of the eleven southern states that had joined the Confederate States of America. Lincoln's re-election ensured that he would preside over the successful conclusion of the Civil War.\n", "These elections, corresponding with Abraham Lincoln's re-election as president, saw the Republicans gain two seats. As these elections occurred during the Civil War, most of the Southern states were absent.\n", "Abraham Lincoln's constitutional election reflected the nation's sectional divide, though 82 percent of the electorate had split among the Unionists, Lincoln, Stephen A. Douglas and John Bell. Even before Lincoln's inauguration, the Deep South states that had cast Electoral College votes for John C. Breckinridge resolved to secede from the United States and form the Confederate States of America. The Virginia Assembly called a special convention for the sole purpose of considering secession from the United States. Virginia was deeply divided, returning a convention of delegates amounting to about one-third for secession and two thirds Unionist. But the Unionists would prove to be further divided between those who would be labelled Conditional Unionists who would favor Virginia in the Union only if Lincoln made no move at \"coercion\", and those who would later be called Unconditional Unionists who would be unwavering in their loyalty to the constitutional government of the United States.\n", "Because eleven Southern states had declared secession from the Union and formed the Confederate States of America, only twenty-five states participated in the election. Three new states participated for the first time: Kansas, West Virginia, and Nevada. The reconstructed portions of Louisiana and Tennessee chose presidential electors - Congress did not count their votes, which could not have changed the result and in any case had been cast for Lincoln. Despite Kentucky's state government never seceding from the Union, the Commonwealth had an election participation rate decrease of almost 40% compared to the election of 1860.\n", "On November 7, 1860, the U.S. presidential election produced a victory for Abraham Lincoln and precipitated the secession of seven slave states and the formation of the Confederate States of America. These events eliminated any chance for federal endorsement of the Territory of Jefferson and any role in government for Governor Steele, a staunch pro-Union Democrat and vocal opponent of Lincoln and the Republican Party.\n" ]
Can somebody explain possible reasons why "Super Luminous Supernova" differ from the garden variety supernova?
I mean... the best source is probably [the paper](_URL_0_) itself. But, **TL;DR** we don't really know yet. It seems too energetic for most models, so the theorists will need some time to adjust. > Only within the past two decades has the most luminous class of supernovae (super-luminous supernovae, SLSNe) been identified. Compared with the most commonly discovered SNe (Type Ia), SLSNe are more luminous by over two magnitudes at peak and rarer by at least 3 orders of magnitude. > The power source for ASASSN-15lh is unknown. Traditional mechanisms invoked for normal SNe likely cannot explain SLSNe-I. The lack of hydrogen or helium suggests that shock interactions with hydrogen-rich circumstellar material, invoked to interpret some SLSNe, cannot explain SLSNe-I or ASASSN-15lh. > Another possibility is that the spindown of a rapidly rotating, highly magnetic neutron star (a magnetar) powers the extraordinary emission. The total observed energy radiated so far (1.1 ± 0.2 × 10^(52) ergs) strains a magnetar interpretation because, for P ≲ 1 ms, gravitational wave radiation should limit the total rotational energy available
[ "Theoretical studies indicate that most supernovae are triggered by one of two basic mechanisms: the sudden re-ignition of nuclear fusion in a degenerate star or the sudden gravitational collapse of a massive star's core. In the first class of events, the object's temperature is raised enough to trigger runaway nuclear fusion, completely disrupting it. Possible causes are accumulation of sufficient material from a binary companion through accretion, or a merger. In the massive star case, the core of a massive star may undergo sudden collapse, releasing gravitational potential energy as a supernova. While some observed supernovae are more complex than these two simplified theories, the astrophysical mechanics have been established and accepted by most astronomers for some time.\n", "Supernovae can expel several solar masses of material at speeds up to several percent of the speed of light. This drives an expanding and fast-moving shock wave into the surrounding interstellar medium, sweeping up an expanding shell of gas and dust observed as a supernova remnant. Supernovae are a major source of elements in the interstellar medium from oxygen through to rubidium. The expanding shock waves of supernovae can trigger the formation of new stars. Supernova remnants might be a major source of cosmic rays. Supernovae might produce strong gravitational waves, though, thus far, the gravitational waves detected have been from the merger of black holes and neutron stars.\n", "Supernovae are the most dramatic type of cataclysmic variable, being some of the most energetic events in the universe. A supernova can briefly emit as much energy as an entire galaxy, brightening by more than 20 magnitudes (over one hundred million times brighter). The supernova explosion is caused by a white dwarf or a star core reaching a certain mass/density limit, the Chandrasekhar limit, causing the object to collapse in a fraction of a second. This collapse \"bounces\" and causes the star to explode and emit this enormous energy quantity. The outer layers of these stars are blown away at speeds of many thousands of kilometers an hour. The expelled matter may form nebulae called \"supernova remnants\". A well-known example of such a nebula is the Crab Nebula, left over from a supernova that was observed in China and North America in 1054. The core of the star or the white dwarf may either become a neutron star (generally a pulsar) or disintegrate completely in the explosion.\n", "Unusually bright supernovae are thought to result from the death of very large stars as pair-instability supernovae (or pulsational pair-instability supernovae). However, recent research by astronomers has postulated that energy released from newly formed magnetars into the surrounding supernova remnants may be responsible for some of the brightest supernovae, such as SN 2005ap and SN 2008es.\n", "Another hypothesis discussed is that effects of a supernova could have been a factor in the Younger Dryas. Effects of a supernova have been suggested before, but without confirming evidence. Potential evidence that these effects could have been caused by a celestial event, a supernova are observations of Gamma-ray bursts and X-ray flashes have been compared to nebular records to test this as well as supernovae flash models, comparable to the records of in-galaxy supernovae, to study the effects of such an event on Earth. These effects include depletion in the ozone layer, increased UV exposure, global cooling, and nitrogen changes in the Earth's surface and troposphere. As Brakenridge states, the only supernova possible at that time was the Vela Supernova, or classified as the Vela Supernova Remnant.\n", "A supernova may also result from mass transfer onto a white dwarf from a star companion in a double star system. The Chandrasekhar limit is surpassed from the infalling matter. The absolute luminosity of this latter type is related to properties of its light curve, so that these supernovae can be used to establish the distance to other galaxies. One of the most studied supernovae is SN 1987A in the Large Magellanic Cloud.\n", "Type Ia supernovae are believed to be thermonuclear explosions that occur when white dwarf stars in binary systems accrete mass from their companion stars. These events are important for the study of cosmology because they are very bright, which allows astronomers to detect them at very large distance. The expansion of the universe can be constrained based on observations of the luminosity distance and redshift of distant type IA supernova. The other three techniques (BAO, galaxy clusters, and weak lensing) used by the Dark Energy Survey allow scientists to understand simultaneously the expansion of the universe and the evolution of the dark matter density field perturbations. These perturbations were intrinsically tied to the formation of galaxies and galaxy clusters. The standard model of cosmology assumes that quantum fluctuations of the density field of the various components that were present when our universe was very young were enhanced through a very rapid expansion called inflation. Gravitational collapse enhances these initial fluctuation as baryons fall into the gravitational potential field of more dense regions of space to form galaxies. Nevertheless, the growth rate of these dark matter halos is sensitive to the dynamics of the expansion of the Universe and DES will use this connection to probe the properties of that expansion.\n" ]
if earth's orbit is elliptical, how is it that summer occurs when earth is furthest from the sun, and spring and fall, when earth is closest to the sun, is cooler than summer?
The change in orbit is more or less negligible. You're talking 3 million miles out of about 91 million miles, about a 3% change. The tilt of the Earth 23.4 degrees. Now think about shining a flash light directly at the wall. It makes a spot right? If you tilt the angle around 30 degrees what happens? The spot stretches out, it covers twice as much ground as it did when you were pointing straight. But the flash light didn't change its power level, the 'spot' is still delivering x energy to the wall, over twice as much area. So any individual 'unit' of that spot is receiving half the energy it was before. While not the only factor, that's a big one. Any given spot on the surface in winter is receiving far less energy than it does in summer.
[ "The Earth's orbit is elliptical, with the Sun at one focus; lines drawn through the summer and winter solstice; and the spring and autumn equinox; intersect with the sun at right angles. The Earth is closest to the Sun (perihelion) near the northern hemisphere winter solstice. The earth moves faster through its orbit when closer to the sun. Hence, the period from the northern hemisphere's autumn equinox to winter and spring is shorter by around seven days than the period from spring to summer to autumn; the reverse is true in the southern hemisphere. Hence, northern hemisphere winter is shorter.\n", "BULLET::::- Seasons are not caused by the Earth being closer to the Sun in the summer than in the winter, but by the Earth's 23.4-degree axial tilt. Each Hemisphere is tilted towards the Sun in its respective summer (July in the Northern Hemisphere and January in the Southern Hemisphere), resulting in longer days and more direct sunlight, with the opposite being true in the winter.\n", "Earth's axis remains tilted in the same direction with reference to the background stars throughout a year (regardless of where it is in its orbit). This means that one pole (and the associated hemisphere of Earth) will be directed away from the Sun at one side of the orbit, and half an orbit later (half a year later) this pole will be directed towards the Sun. This is the cause of Earth's seasons. Summer occurs in the Northern hemisphere when the north pole is directed toward the Sun. Variations in Earth's axial tilt can influence the seasons and is likely a factor in long-term climate change \"(also see Milankovitch cycles)\".\n", "Because of the increased distance at aphelion, only 93.55% of the solar radiation from the Sun falls on a given area of land as does at perihelion. However, this fluctuation does not account for the seasons, as it is summer in the northern hemisphere when it is winter in the southern hemisphere and \"vice versa.\" Instead, seasons result from the tilt of Earth's axis, which is 23.4 degrees away from perpendicular to the plane of Earth's orbit around the sun. Winter falls on the hemisphere where sunlight strikes least directly, and summer falls where sunlight strikes most directly, regardless of the Earth's distance from the Sun. In the northern hemisphere, summer occurs at the same time as aphelion. Despite this, there are larger land masses in the northern hemisphere, which are easier to heat than the seas. Consequently, summers are warmer in the northern hemisphere than in the southern hemisphere under similar conditions. Astronomers commonly express the timing of perihelion relative to the vernal equinox not in terms of days and hours, but rather as an angle of orbital displacement, the so-called longitude of the periapsis (also called longitude of the pericenter). For the orbit of the Earth, this is called the \"longitude of perihelion\", and in 2000 it was about 282.895°; by the year 2010, this had advanced by a small fraction of a degree to about 283.067°.\n", "Over thousands of years, the Earth's axial tilt and orbital eccentricity vary (see Milankovitch cycles). The equinoxes and solstices move westward relative to the stars while the perihelion and aphelion move eastward. Thus, ten thousand years from now Earth's northern winter will occur at aphelion and northern summer at perihelion. The severity of seasonal change — the average temperature difference between summer and winter in location — will also change over time because the Earth's axial tilt fluctuates between 22.1 and 24.5 degrees.\n", "The Sun appears to move northward during the northern spring, contacting the celestial equator on the March equinox. Its declination reaches a maximum equal to the angle of Earth's axial tilt (23.44°) on the June solstice, then decreases until reaching its minimum (−23.44°) on the December solstice, when its value is the negative of the axial tilt. This variation produces the seasons.\n", "Warm summers in the Southern hemisphere occur when that hemisphere is tilted toward the sun and the Earth is nearest the sun in its elliptical orbit. Cool summers occur when the Earth is farthest from the sun during that season. These effects are more pronounced when the eccentricity of the orbit is large. When the obliquity is large, seasonal changes are more extreme.\n" ]
-why is that if you get distracted, your muscles become weaker?
Grip strength is a voluntary muscle movement, therefore by distracting them you are taking their mind off that action and so the signal to it will be decreased.
[ "When concentrating on a visually intense task, such as continuously focusing on a book or computer monitor, the ciliary muscle tightens. This can cause the eyes to get irritated and uncomfortable. Giving the eyes a chance to focus on a distant object at least once an hour usually alleviates the problem.\n", "Early experiments showed that adrenaline increases twitch, but not tetanic force and rate of force development in muscles. It is questionable, however, as to whether adrenaline, released from the adrenal medulla into the venous circulation, can reach the muscle quickly enough in order to be able to cause such an effect in the midst of a crisis. It may be that noradrenaline released from sympathetic nerve terminals directly innervating skeletal muscle has more of an effect over the timescale of seconds.\n", "Distraction, an example of attentional deployment, is an early selection strategy, which involves diverting one's attention away from an emotional stimulus and towards other content. Distraction has been shown to reduce the intensity of painful and emotional experiences, to decrease facial responding and neural activation in the amygdala associated with emotion, as well as to alleviate emotional distress. As opposed to reappraisal, individuals show a relative preference to engage in distraction when facing stimuli of high negative emotional intensity. This is because distraction easily filters out high-intensity emotional content, which would otherwise be relatively difficult to appraise and process.\n", "In novice strength trainers, the muscle's ability to generate force is most strongly limited by nerve’s ability to sustain a high-frequency signal. After a period of maximum contraction, the nerve’s signal reduces in frequency and the force generated by the contraction diminishes. There is no sensation of pain or discomfort, the muscle appears to simply ‘stop listening’ and gradually cease to move, often going backwards. As there is insufficient stress on the muscles and tendons, there will often be no delayed onset muscle soreness following the workout.\n", "For extremely powerful contractions that are close to the upper limit of a muscle's ability to generate force, neuromuscular fatigue can become a limiting factor in untrained individuals. In novice strength trainers, the muscle's ability to generate force is most strongly limited by nerve’s ability to sustain a high-frequency signal. After an extended period of maximum contraction, the nerve’s signal reduces in frequency and the force generated by the contraction diminishes. There is no sensation of pain or discomfort, the muscle appears to simply ‘stop listening’ and gradually cease to move, often lengthening. As there is insufficient stress on the muscles and tendons, there will often be no delayed onset muscle soreness following the workout. Part of the process of strength training is increasing the nerve's ability to generate sustained, high frequency signals which allow a muscle to contract with their greatest force. It is this \"neural training\" that causes several weeks worth of rapid gains in strength, which level off once the nerve is generating maximum contractions and the muscle reaches its physiological limit. Past this point, training effects increase muscular strength through myofibrillar or sarcoplasmic hypertrophy and metabolic fatigue becomes the factor limiting contractile force.\n", "For extremely powerful contractions that are close to the upper limit of a muscle's ability to generate force, neuromuscular fatigue can become a limiting factor in untrained individuals. In novice strength trainers, the muscle's ability to generate force is most strongly limited by nerve’s ability to sustain a high-frequency signal. After an extended period of maximum contraction, the nerve’s signal reduces in frequency and the force generated by the contraction diminishes. There is no sensation of pain or discomfort, the muscle appears to simply ‘stop listening’ and gradually cease to move, often lengthening. As there is insufficient stress on the muscles and tendons, there will often be no delayed onset muscle soreness following the workout. Part of the process of strength training is increasing the nerve's ability to generate sustained, high frequency signals which allow a muscle to contract with their greatest force. It is this \"neural training\" that causes several weeks worth of rapid gains in strength, which level off once the nerve is generating maximum contractions and the muscle reaches its physiological limit. Past this point, training effects increase muscular strength through myofibrillar or sarcoplasmic hypertrophy and metabolic fatigue becomes the factor limiting contractile force.\n", "As a result of this effect, not only is the soreness reduced, but other indicators of muscle damage, such as swelling, reduced strength and reduced range of motion, are also more quickly recovered from. The effect is mostly, but not wholly, specific to the exercised muscle: experiments have shown that some of the protective effect is also conferred on other muscles.\n" ]
What unified the Austrian-Hungary Empire - Religion, Culture, etc.?
It was conquest! Always conquest. The Austrian Empire, or the Austro-Hungarian Empire, had its roots with the Habsburgs, and the Duchy of Austria. Austria and Vienna lay on trade roots coming east out of Italy, North out of the Balkans, and South and East out of the Baltics/Ukrainian steppe. Throughout the Renaissance and Early Modern period, this made Austria wealthy and powerful. The Habsburgs, Austria's ruling family, also gained tremendous political power across most of Germany, they became the Emperor of all of Germany. Austria used this power and prestige to spread their rule across southern and eastern Germany. They joined the thrones of Bohemia and Hungary to the Austrian one, giving Austria de facto rule over the area, while northern Italy was drug into Austria through a series of wars, and an advantageous marriage in Spain, which for a time unified the Austrian and Spanish thrones. But all this power, wealth, and prestige also made Austria a target. It waged several notable campaigns against the Ottoman Turks, who after toppling the Byzantines continued to push north through the Balkans. Austria continued to strengthen its southern border, as well as Hungary's southern border, to prevent [serious Turkish invasions](_URL_0_). Thus Austria began to see itself as the traditional defender of Christianity in the Balkans, as well as the defender of Europe against the Islamic Turks. So really, Austria, and later Austro-Hungary, was a polyglot empire because Austria expanded in such a haphazard and seemingly random way. The most important connecting thread was always "what is best for the Habsburgs?" who sat at the top of the whole pyramid. In the Renaissance, it made Austria a dynamic political and military power, which hung a curtain across southern Germany. But following the Thirty Years Wars, and especially after the Napoleonic Wars, the multitude of nationalities, religions, political and social ideas, and ethnic alignments all made Austria a strange and unhealthy nation. The Habsburgs continued to rule as if the Austrian Empire were still the "Habsburg Lands", which you sometimes see printed on older maps of the Renaissance. This led to the breakdown of Austria, which really culminated in the July Crisis of 1914, which directly caused the conflict which would destroy the Austrian (at that point Austro-Hungarian) Empire in 1918.
[ "The \"Austro-Hungarian Compromise of 1867\" created the personal union of the independent states of Hungary and Austria, linked under a common monarch also having joint institutions. The Hungarian majority asserted more of their identity within the Kingdom of Hungary, and it came to conflict with some of her own minorities. The imperial power of German speakers who controlled the Austrian half was resented by others. In addition, the emergence of nationalism in the newly independent Romania and Serbia also contributed to ethnic issues in the empire.\n", "Because of external and internal problems, reforms seemed inevitable and major military defeats of Austria forced the Habsburgs to negotiate the Austro-Hungarian Compromise of 1867, by which the dual Monarchy of Austria–Hungary was formed. This Empire had the second largest area in Europe (after the Russian Empire), and it was the third most populous (after Russia and the German Empire). The two realms were governed separately by two parliaments from two capital cities, with a common monarch and common external and military policies. Economically, the empire was a customs union. The old Hungarian Constitution was restored, and Franz Joseph I was crowned as King of Hungary. The era witnessed impressive economic development. The formerly backward Hungarian economy became relatively modern and industrialized by the turn of the 20th century, although agriculture remained dominant until 1890. In 1873, the old capital Buda and Óbuda were officially united with Pest, thus creating the new metropolis of Budapest. Many of the state institutions and the modern administrative system of Hungary were established during this period.\n", "The Austro-Hungarian monarchical personal union of the two countries was a result of the Compromise of 1867. Kingdom of Hungary lost its former status after the Hungarian Revolution of 1848. However following the 1867 reforms, the Austrian and the Hungarian states became co-equal within the Empire. Austria-Hungary was geographically the second-largest country in Europe after the Russian Empire, at , and the third-most populous (after Russia and the German Empire). In comparison with Germany and Britain, the Austro-Hungarian economy lagged behind considerably, as sustained modernization had begun much later in Austria-Hungary. The Empire built up the fourth-largest machine building industry of the world, after the United States, Germany, and Britain. Austria-Hungary was also the world's third largest manufacturer and exporter of electric home appliances, electric industrial appliances and facilities for power plants, after the United States and the German Empire.\n", "Austria-Hungary, often referred to as the Austro-Hungarian Empire or the Dual Monarchy, was a constitutional monarchy in Central and Eastern Europe between 1867 and 1918. It was formed when the Austrian Empire adopted a new constitution; as a result Austria (Cisleithania) and Hungary (Transleithania) were placed on equal footing. It dissolved into several new states at the end of the First World War.\n", "Austria-Hungary was a monarchic union between the crowns of the Austrian Empire and the Kingdom of Hungary in Central Europe. The union was a result of the \"Ausgleich\" or Compromise of 1867, under which the Austrian House of Habsburg agreed to share power with the separate Hungarian government, dividing the territory of the former Austrian Empire between them. The Dual Monarchy had existed for 51 years when it dissolved on October 31, 1918 following military defeat in the First World War.\n", "The Austro-Hungarian Empire, headquartered at Vienna, was a largely rural, poor, multicultural state. It was operated by and for the Habsburg family, who demanded loyalty to the throne, but not to the nation. Nationalistic movements were growing rapidly. The most powerful were the Hungarians, who preserved their separate status within the Habsburg Monarchy and with the Austro-Hungarian Compromise of 1867, the creation of the Dual Monarchy they were getting practical equality. Other minorities, were highly frustrated, although some – especially the Jews – felt protected by the Empire. German nationalists, especially in the Sudetenland (part of Bohemia) however, looked to Berlin in the new German Empire. There was a small German-speaking Austrian element located around Vienna, but it did not display much sense of Austrian nationalism. That is it did not demand an independent state, rather it flourished by holding most of the high military and diplomatic offices in the Empire. Russia was the main enemy, As well as Slavic and nationalist groups inside the Empire (especially in Bosnia-Herzegovina) and in nearby Serbia. Although Austria, Germany, and Italy had a defensive military alliance – the Triple Alliance – Italy was dissatisfied and wanted a slice of territory controlled by Vienna. \n", "After years of dealing with Hungarian revolutionist, whose kingdom Austria had conquered centuries earlier, the Austrian emperor, Franz Joseph agreed to divide the empire into two parts: Austria and Hungary, and rule as both Emperor of Austria and king of Hungary. The new Austro-Hungarian Empire was created in 1867. The two peoples were united in loyalty to the monarch and Catholicism.\n" ]
why is it so much easier to spend money than to earn it?
Because most employment will only pay you a set amount for an hour's work, limiting you in how much money you can make in a span of time, but you can never exhaust the human race's output of purchasable goods.
[ "It has been argued that money cannot effectively \"buy\" much happiness unless it is used in certain ways, and that \"Beyond the point at which people have enough to comfortably feed, clothe, and house themselves, having more money – even a lot more money – makes them only a little bit happier.\" In his book \"Stumbling on Happiness\", psychologist Daniel Gilbert described research suggesting money makes a significant difference to the poor (where basic needs are not yet met), but has a greatly diminished effect once one reaches middle class (i.e. the Easterlin paradox). Every dollar earned is just as valuable to happiness up to a $75,000 annual income, thereafter, the value of each additional dollar earns a diminishing amount of happiness. According to the latest systematic review of the economic literature on life satisfaction, one's perception of their financial circumstances fully mediates the effects of objective circumstances on one's well-being. People overestimate the influence of wealth by 100%.\n", "Indeed, when one has met his basic needs and have more to spare, it is time to spend or give some to experience happiness. This is because happiness is really a state of in-and-out flow of one's energy. Using or giving money is an expression of out-flowing of one's life-state. Attempt to just hoard more and more in the belief that it brings more happiness can lead to the opposite result if only because the means – that is the pursuit of money for happiness – has unwittingly become the ends.\n", "Some studies suggest, however, that people are happier after spending money on experiences, rather than physical things, and after spending money on others, rather than themselves. However, purchases that buy ‘time’, for instance, cleaners or cooks typically increase individual well-being.\n", "Other things being equal, the greater the amount of profit-income that is disbursed as personal earnings and used for consumptive purposes, the lower the savings rate and the lower the rate of accumulation is likely to be. However, earnings spent on consumption can also stimulate market demand and higher investment. This is the cause of endless controversies in economic theory about \"how much to spend, and how much to save\".\n", "A third is that there is a \"persistent and disturbing reluctance of businesses to invest and consumers to spend\", perhaps in part because so much of the recent gains have gone to the people at the top, and they tend to save more of their money than people—ordinary working people who can't afford to do that.\n", "A third is that there is a \"persistent and disturbing reluctance of businesses to invest and consumers to spend\", perhaps in part because so much of the recent gains have gone to the people at the top, and they tend to save more of their money than people—ordinary working people who can't afford to do that.\n", "Some people practice simple living by reducing consumption. By lowering expenditure on goods or services, the time spent earning money can be reduced. The time saved may be used to pursue other interests, or help others through volunteering. Some may use the extra free time to improve their quality of life, for example pursuing creative activities such as art and crafts. Developing a detachment from money has led some individuals, such as Suelo and Mark Boyle, to live with no money at all. Reducing expenses may also lead to increasing savings, which can lead to financial independence and the possibility of early retirement.\n" ]
What is the process for a new atom or element to form, specifically from the beginning when there was only hydrogen and helium?
> If all matter began from hydrogen and helium, how did we end up with 120+ elements? There are 118 known elements, and some of them don't occur in significant amounts in nature. > Is it possible to create a specific element by mashing x amount of protons, neutrons, and electrons together? Yes, although there are typically easier ways of producing a given element. We have many different kinds of nuclear reactions in our arsenal, and many stable (or nearly-stable) nuclides that we can use as a starting point. > Obviously I know this is not how it works AT ALL but how could other elements form from just 2 elements? [Here](_URL_0_) is a chart of most of the currently-known elements, with the primary production mechanisms shown. In these astrophysical sites, like neutron star mergers and supernovae, there are complicated networks of many nuclear reactions and decays happening. They produce many different isotopes of many different elements.
[ "Since there had been no previous star formation to create other elements, protogalaxies would have been made up almost entirely of hydrogen and helium. The hydrogen would bond to form H molecules, with some exceptions. This would change as star formation began and produced more elements through the process of nuclear fusion.\n", "At around 100,000 years, the universe has cooled enough for helium hydride, the first molecule, to form. In April 2019, this molecule was first announced to have been discovered in interstellar space. (Much later, atomic hydrogen reacts with helium hydride to create molecular hydrogen, the fuel required for star formation.)\n", "At around 377,000 years, the universe has cooled to a point where free electrons can combine with the hydrogen and helium nuclei to form neutral atoms. This process is relatively fast (and faster for the helium than for the hydrogen), and is known as recombination. The name is slightly inaccurate and is given for historical reasons: in fact the electrons and atomic nuclei were combining for the first time.\n", "Since the Big Bang, which produced no carbon or heavier elements, atomic nuclei have been combined in stars through the process of nuclear fusion to produce more of the element helium, and (via the triple alpha process) the sequence of elements from carbon up to iron; see stellar nucleosynthesis for details.\n", "In this form of decay, the original element becomes a new chemical element in a process known as nuclear transmutation. This new element has an unchanged mass number , but an atomic number that is increased by one. As in all nuclear decays, the decaying element (in this case ) is known as the \"parent nuclide\" while the resulting element (in this case ) is known as the \"daughter nuclide\".\n", "All atoms bigger than hydrogen are formed in stars or supernovae through nucleosynthesis, when gravity, temperature and pressure reach levels high enough to fuse protons and neutrons together. Protons and neutrons form the atomic nucleus, which accumulates electrons to form atoms. The number of protons in the nucleus, called atomic number, uniquely identifies a chemical element.\n", "Albert Einstein's development of the theory of relativity in 1905 led to the understanding that nuclear reactions could create new elements from smaller precursors, with the loss of energy. In his treatise \"Stars and Atoms\", Arthur Eddington suggested that pressures and temperatures within stars were great enough for hydrogen nuclei to fuse into helium; a process which could produce the massive amounts of energy required to power the Sun. In 1935, Eddington went further and suggested that other elements might also form within stars. Spectral evidence collected after 1945 showed that the distribution of the commonest chemical elements, carbon, hydrogen, oxygen, nitrogen, neon, iron etc., was fairly uniform across the galaxy. This suggested that these elements had a common origin. A number of anomalies in the proportions hinted at an underlying mechanism for creation. Lead has a higher atomic weight than gold, but is far more common. Hydrogen and helium (elements 1 and 2) are virtually ubiquitous yet lithium and beryllium (elements 3 and 4) are extremely rare.\n" ]
why do we try to keep people in vegetative states alive, who are not going to recover?
There's several reasons. It's the medical team's ethical and professional duty to maintain the person's life; they have a duty of care to the patient. It could be considered medical negligent or even murder/manslaughter if they ended the person's life. Also, they can't say for sure whether the person will recover or not.
[ "3. Most people want to be kept \"alive\" by machines. Most people don't want to drain their family's funds to keep them alive especially when they are in a persistent vegetative state with no possible chance for recovery. In these cases, it is often weighing the risks and benefits of keeping the patient breathing, when they are clearly not living their life to its full potential.\n", "The revivers are immortal and heal from all wounds. Some of them begin to take physical and non-physical risks because they do not fear physical or emotional harm to themselves or others. When they experience strong negative emotions, they cry blood and become violent. As a result, Dana and Ramin investigate several murders in the weeks following Revival Day. Meanwhile, some people outside the quarantine area believe the government is covering up a religious miracle. Others believe they can absorb the revivers' immortality by ingesting their flesh, leading to an active smuggling business that moves body parts of revivers and other recently dead individuals.\n", "No two people respond to resocialization programs in the same manner. While some residents are found to be \"rehabilitated\", others might become bitter and hostile. As well, over a long period of time, a strictly controlled environment can destroy a person's ability to make decisions and live independently. This is known as institutionalisation, a negative outcome of total institution that prevents an individual from ever functioning effectively in the outside world again. (Sproule, 154-155)\n", "Apart from using hospitals and healthcare services to regain their health if it has become impaired, or to prevent ill health occurring in the first place, people the world over may also use them for a wide variety of other services, for example “improving upon nature” (e.g. cosmetic surgery, gender reassignment surgery or acquiring help to overcome difficulties with becoming a parent (e.g. infertility treatment).\n", "People who have survived conflict or a natural disaster often have acute needs. Frequently, they have been displaced from their homes and lack their usual economic, social or psychological support systems.\n", "Those left alive must continue to cope with their problems. Knowing some answers does not really resolve the divisions and imperfections in society which it is the job of Rebus and his colleagues to police.\n", "Today in many countries there is a shortage of hospital space. Medical personnel and hospital beds could be used for people whose lives could be saved instead of continuing the lives of those who want to die, thus increasing the general quality of care and shortening hospital waiting lists. It is a burden to keep people alive past the point they can contribute to society, especially if the resources used could be spent on a curable ailment.\n" ]