text
stringlengths
5
1.05M
The slave trade, yet horrific in it’s inhumanity, became an important aspect of the world’s economy during the eighteenth century. During a time when thousands of Africans were being traded for currency, Olaudah Equiano became one of countless children kidnapped and sold on the black market as a slave. Slavery existed centuries before the birth of Equiano (1745), but strengthened drastically due to an increasing demand for labor in the developing western hemisphere, especially in the Caribbean and Carolinas. Through illogical justification, slave trading became a powerful facet of commerce, regardless of its deliberate mistreatment of human beings by other human beings. Olaudah Equiano was able to overcome this intense adversity and actually accumulate wealth by making the best of certain situations he faced throughout his experiences. Even though he was a victim of the slave trade, he willfully took advantage of the opportunity to see the world and to become a productive individual. Olaudah Equiano and his sister were kidnapped by slave traders from their native village in Isseke, Nigeria. The nature of the African village, Isseke, was labor based and emphasized the concept of earning profit from hard work. There was a strong sense of community among these villagers, as Equiano explains in his personal narrative, “Thus we are all habituated to labor from our earliest years. Everyone contributes something to the common stock; and, as we are unacquainted with idleness, we have no beggars” (39) The concept of slavery was accepted as a part of the culture and even in the fields of Isseke, Africa slaves were put to work. “Sometimes indeed, we sold slaves to them, but they were only prisoners of war, or such among us as had been convicted of kidnapping or adultery, and some other crimes, which we esteemed heinous.” (38) Equiano’s fortune landed him in the hands of a wealthy widow who purchased him from the traders who had kidnapped him. He lived the life as a companion to the widow and her son. Luck was on his side in this transaction, many slave owners frowned upon educating and assisting slaves. “Masters” typically feared an educated slave would take measures to make a change. He explains, though, how he held status above other slave under the widow’s ownership, “There were likewise slaves daily to attend us, while my young master and I, with other boys, sported with our darts and bows and arrows, as I had been used to at home. ” (52) Equiano, a slave at the time, had the shared the power of a slave master! Equiano’s luck soon shifted when he was once again kidnapped and sold as a slave, this time he would have to endure the notoriously dreadful journey across the sea to America. Knowing that this was a pivotal point in his life and that he would become a gudgeon to the harshness of slavery, Equiano attempted to prepare himself for what lay ahead. However, the sight of the inhumane acts he witnessed on the African coast, while being transported, were new to Equiano and instilled fear into his consciousness. Equiano, on a slave ship towards the West Indies, was on the verge of everlasting bondage. “In this manner we continued to undergo more hardships than I can now relate, hardships which are inseparable from this accursed trade. Many a time we were near suffocation from the want of fresh air, which was often without for whole days together.” (57) The conditions for slaves on these transport ships was horrendous, as those in charge cared little for the well being of these Africans. Equiano was unaware of what was to come, and fear lingered in his memory of this unforgiving experience. He explains the process of the transaction, “We were not many days in the merchant’s custody, before we were sold after their usual manner, which was: On a signal given (as the beat of a drum), the buyers rush at once into the yard where the slaves are confined, and make choice of that parcel they like best.” (58) Subsequent to being purchased at the slave auction in 1763, Equiano began his journey as a slave in the New World and was stationed on the island of Montserrat in the Caribbean. After being traded as a slave from master to master for some time, he came to serve for a lieutenant in the Royal Navy. Following the Revolutionary war, Equiano pursued to better his situation throughout his travels, and continued to learn as much as he could while under the serfdom of Admiral Bosca wen. He continued to interpret the ethics of slavery and made efforts to change the way slave master’s treated their slaves. “But by changing your conduct, and treating your slaves as men, every cause of fear would be banished.” He explains in the name of his fellow slaves, “They would be faithful, honest, intelligent, and vigorous; and peace, prosperity, and happiness would attend you.” (100) The idea of becoming free was the ultimate dream for Equiano. When he earned money, he used it for his education. “I therefore employed the mate of our vessel to teach me navigation, for which I agreed to give him twenty-four dollars, and actually paid him part of the money down” (108) Navigation, an important and useful skill, would also help Equiano become a will balanced individual in society. “I had long wished to be able to read and write; and for this purpose I took every opportunity to gain instruction, but had made very little progress. However, when I went to London with my master, I had soon an opportunity of improving myself, which I gladly embraced. Shortly after my arrival, he sent me to wait upon Miss Guerin’s, who had treated me with much kindness when I was there before; and they sent me to school.” (72) Equiano eventually saved enough money to buy his freedom. This was a rare occurrence, for most slaves did not earn nearly enough money to do such a thing. But Equiano developed skills that allowed him to hustle the system and became a free man. With new found freedom, he set forth to apply it where he could, as an activist. “From that period, to the present time, my life has passed in an even tenor, and a great part of my study and attention has been to assist in the cause of much injured countrymen.” (191) Equiano became active by promoting and petitioning slavery and dedicated his life to the freedom of his enslaved people. Racial adversity has proven very difficult to overcome, even in today’s society. Olaudah Equiano used his economic intelligence to provide him with the power to accumulate wealth and buy his freedom. Equiano knew he was dealt an uneasy hand of cards at life, but he played them as skillfully and intelligently as he could, with bare instincts. He took advantage of the situations that he was forced into and was able to win his hand at the game of slavery, for he found a way to manipulate the system that opposed him, and used it to his advantage. Works cited The Interesting Narrative of the life of Olaudah Equiano: written by Himself.
Declaration of the Party of the European Left on Lebanon Maite Mola, Vice-President of the European Left: The Party of the European Left expresses its deep concern after the events that occurred since November 4 with the resignation of Prime Minister Saad Hariri, in particularly turbulent conditions, and contrary to the constitutional rules, as pointed out by the highest authorities of the Lebanese State. This situation comes at a time when Lebanon had regained a remarkable stability over the last year, marked in particular by the return to the functioning of the institutions, the election of the President of the Republic, the formation of a government of national unity, the adoption of the Finance Act and a new electoral law introducing for the first time the principle of proportional representation in the legislative elections. At the same time, the terrorist groups Daesh and Al-Nosra had been weakened, allowing for greater stability and security. In view of the development of this crisis as unexpected as it is serious, the Party of the European Left: Calls for full clarity on the conditions under which Lebanese Prime Minister Saad Hariri is kept out of the national territory in detention conditions which are becoming particularly strange and suspect from day to day. Supports the President of the Lebanese Republic, General Michel Aoun in his demand for the return of Prime Minister Saad Hariri and his family to his homeland as soon as possible. Assures, like the Lebanese authorities, that this is the express condition for the return to the constitutional life of the institutions and the easing of tensions. Supports the position of the authorities and the Government of Lebanon on the respect of the rules and the foundations of international law, including the principles of sovereignty, non-interference and independence enshrined in the United Nations Charter, the Vienna Convention and the founding charter of the Arab League. The EL calls Saudi-Arabia and Iran to stop intervening in Lebanese domestic affairs. Calls on all States of the European Union to strongly support the common position expressed by the EU and to support Lebanon in this event. Calls on the European Parliament, its President, as well as all relevant parliamentary bodies engaged with Lebanon in the framework of the Association Agreements to take action and implement all means to help Lebanon and strengthen its stability and its development.
How to use the quick design function in eCommerceA guide on what changes can be made to the look of your eCommerce online shop from the quick design interface. Click here to download the Using quick design in eCommerce guide. What will this guide help me with? This guide contains information on the following areas: - Backing up your template. - Adding and editing images on your pages. - Changing text Fonts and Colours. - Changing the icon set.
हैं जिन्होंने एनआरसी सीएए प्रोटेस्ट के दौरान लोगों को हिंसा में शामिल होने के लिए भड़काया था. इसके प्रोड्यूसर फ़रहान अख़्तर ने रिया (चक्रवर्ती), जया बच्चन और ज़ाहिर सी बात है कि सीएए का भी समर्थन किया. शरीर को ऊर्जावान रखने के लिए योग अफ्रीकी देश जिम्बाब्वे के पूर्व तानाशाह और 37 साल तक देश के राष्ट्रपति रहे रॉबर्ट मुगाबे का शुक्रवार सुबह निधन हो गया। नारायणपुर। छत्तीसगढ़ के नारायणपुर जिले में बुधवार को पुलिस के साथ मुठभेड़ में दो नक्सलियों मार दिए गए। नारायणपुर एएसपी नीलकंठ साहू ने बताया कि छोटे डोंगर थाना क्षेत्र से बुधवार की सुबह पुलिस का संयुक्त बल गश्त के लिए रवाना किया गया था। लौटते वक्त मडोनार के जंगल में घात लगाए नक्सलियों ने पुलिस पर हमला कर दिया। जवाबी कार्रवाई में पुलिस बल ने भी फौरन मोर्चा संभालते हुए डटकर मुकाबला किया। लगभग एक घंटे की मुठभेड़ के बाद आखिरकार नक्सलियों के पैर उखड़ गए और वे जंगल में भाग खड़े हुए। मधुमक्खियों के लिए न्यूट्रीशियन और लर्निग प्रोग्राम का आयोजन किया जाता था. साथ ही कई तरह की पहेली का प्रयोग किया गया. जब मधुमक्खी पहेली के पास जाती थी तो वहां पर 1 से लेकर 5 तक के नम्बर के आकार</s>
Writing a case study is not like a walk in the park. With this, many students dislike it and other forms of coursework like thesis, dissertation, and long essays. Even if you do not like coursework, you should endure it and thrive because, at the end of the day, the result will reflect your understanding of the course or training. If you did well, you will be rewarded with high marks. It is good if you can cope up with all your coursework but what if you cannot handle all at the same time and the pending tasks just keep on piling? This is the time that you consider coursework writing services but how to proceed if you do my case study? Before anything, here are the things that you need to know about coursework writing services: Advantages of hiring coursework writing services The advantages of hiring coursework writing services include the following: - You are assured that all are quality works because of the professional writers. With professionals, your paper will convey your message correctly without spelling and grammar mistakes. - Aside from the spelling and grammar mistakes, coursework writing services also ensure that case studies, thesis and dissertations are original. This means it is free from plagiarism. - The tasks have fixed submission so you can rest easy knowing that professional writers will submit them on time. Disadvantages of hiring coursework writing services The disadvantages of hiring coursework writing services include the following: - There are those who believe that this impedes the learning process of students. - If you do not get in touch with the writer regularly, the coursework might turn out to be different from what is expected of you and this will cost you more at the end of the day. How to find the right coursework writing services To find the right coursework writing services, you should do the following things: - Do your research: there are many pseudo-professionals around so you have to be wary. The best measure here is to do your research and find about the reputation of the company. You can start with online reviews. You can also ask other friends for recommendations. - Avoid rushing: even if the project is urgent, you should put careful thought into your decision. If you settle for the cheapest option without considering others, you might end up with a poor-quality work. - Ensure that the writer can handle any job: you have to ask the company if they have versatile writers. This means that their writers can handle any type of job. - 24/7 customer service: you should also look into customer support. At the bare minimum, it should be 24/7 with many platforms for reaching out like live chat and voice support. After knowing basic information about coursework writing services, you can now get started by signing up to create an account – just make sure that you fill up all the required fields. When inside, you should explore the menu and look for the option “Write my paper”. Make sure to put the necessary information, choose a writer and then wait for its completion.
10 Awesome Printable Map Indonesia – A map is a symbolic depiction highlighting relationships amongst parts of some distance, such as objects, places, or themes. Most maps are somewhat still static, adjusted to paper or any other durable medium, while some are somewhat interactive or dynamic. Even though most widely utilized to depict geography, maps may reflect some distance, fictional or real, with no respect to scale or circumstance, such as in mind mapping, DNA mapping, or computer system topology mapping. The distance being mapped may be two dimensional, like the surface of the planet, 3d, such as the interior of the planet, or even even more abstract spaces of any measurement, such as for instance arise in simulating phenomena with many independent variables. Even though first maps understood are of the heavens, geographic maps of territory possess a very long tradition and exist from early days. The term”map” comes from the medieval Latin Mappa mundi, wherein mappa meant napkin or cloth along with mundi the world. So,”map” grew to become the abbreviated term speaking into a two-dimensional representation of this surface of the planet. Road maps are perhaps the most commonly used maps today, and also sort a subset of specific maps, which likewise include aeronautical and nautical charts, railroad system maps, and hiking and bicycling maps. With respect to quantity, the biggest quantity of drawn map sheets is probably made up by local polls, completed by municipalities, utilities, tax assessors, emergency services suppliers, as well as also other native organizations. World Map Showing Indonesia Map City Map Showing Canada Us Borders Best Us Canada Map New I Pinimg, Source : windsurfaddicts.com World Map Showing Indonesia Map City Us Canada Map New Uploadmedia Mons 0 0d World Map North, Source : windsurfaddicts.com World Map Showing Indonesia Best Map East Asia The Countries Are China Russia Japan North, Source : windsurfaddicts.com World Map Showing Indonesia New Simple World Map With Country Names Valid Code Golf, Source : windsurfaddicts.com A number of national surveying projects are carried from the armed forces, like the British Ordnance Survey: a civilian government agency, globally renowned because of its detailed operate. Along with position info maps may likewise be used to portray contour lines indicating constant values of altitude, temperature, rainfall, etc.
Unlock the Power of Simulation with MRC’s Advanced Manufacturing Technology Solutions. In the dynamic world of manufacturing, making informed decisions is crucial for success. That’s why MRC is proud to offer Simulation as part of our comprehensive suite of Advanced Manufacturing Technology solutions. Through our strategic partnerships, we provide cutting-edge simulation software that enables the creation of Line Models or Digital Twins, revolutionizing the way manufacturers plan and optimize their operations. Create Line Models or Digital Twins for Enhanced Decision-Making With MRC’s Simulation solutions, manufacturers can create accurate and detailed Line Models or Digital Twins of their production processes. These virtual representations provide a realistic and dynamic view of the entire manufacturing line, allowing you to visualize, analyze, and optimize various scenarios. Whether you’re considering new, large capital investments or reconfiguring existing equipment, our simulation software offers invaluable insights to support your decision-making process. Optimize New Investments with Confidence Investing in new equipment or expanding your manufacturing capabilities can be a significant undertaking. MRC’s Simulation solutions offer a cost-effective and risk-reducing approach by allowing you to virtually test and optimize different production layouts, workflows, and resource allocations before making substantial financial commitments. With our advanced simulation software, you can explore various scenarios, analyze performance indicators, and confidently make data-driven decisions that align with your business goals. Maximize Efficiency through Equipment Reconfiguration As manufacturing needs evolve, reconfiguring existing equipment becomes necessary to accommodate changing demands. MRC’s Simulation solutions empower you to assess the impact of equipment reconfiguration on productivity, cycle times, and resource utilization. By digitally modeling your production processes, you can identify bottlenecks, streamline workflows, and optimize layouts, all without disrupting your actual operations. This enables you to make informed decisions backed by data, leading to increased efficiency and reduced downtime. Partner with MRC for Simulation Excellence At MRC, we understand the importance of leveraging simulation to drive operational excellence. Our team of experts works closely with you to understand your specific requirements and tailor simulation solutions that address your unique manufacturing challenges. Through our partnerships with leading simulation software providers, we offer state-of-the-art tools and comprehensive support to ensure you derive maximum value from your simulation projects. Experience the Power of Simulation with MRC Discover how Simulation can revolutionize your manufacturing operations and drive your business forward. Contact MRC today to explore our Advanced Manufacturing Technology solutions and unlock the potential of Line Models and Digital Twins. Together, we will enhance your decision-making capabilities, optimize your investments, and propel your manufacturing processes to new heights of efficiency and profitability.
Current research suggests conflicting evidence surrounding the association between serum magnesium levels and the diagnosis of attention deficit hyperactivity disorder (ADHD). This systematic review and meta-analysis aims to explore, summarize and quantify the published literature addressing this topic. We conducted an exhaustive literature search on Scopus and PubMed for all the relevant observational studies published up to August 2018. A meta-analysis using a random-effects model was used to summarize the overall association between serum magnesium level and ADHD from the available data. We identified seven studies which reported the mean and standard deviation (SD) of magnesium concentration in both ADHD and control groups. The random-effects meta-analysis showed that subjects with ADHD had 0.105 mmol/l (95% CI: -0.188, -0.022; P < 0.013) lower serum magnesium levels compared with to their healthy controls. Moreover, we observed striking and statistically significant heterogeneity among the included studies (I2 = 96.2%, P = 0.0103). The evidence from this meta-analysis supports the theory that an inverse relationship between serum magnesium deficiency and ADHD exists. High heterogeneity amongst the included studies suggests that there is a residual need for observational and community-based studies to further investigate this issue. Medical opinion and guidance should always be sought for any symptoms that might possibly reflect a known or suspected disease, disorder or medical condition. Information provided on this website (or by FAB Research via any other means) does not in any way constitute advice on the treatment of any medical condition formally diagnosed or otherwise.
# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from __future__ import print_function import sys from random import Random from pyspark import SparkContext numEdges = 200 numVertices = 100 rand = Random(42) def generateGraph(): edges = set() while len(edges) < numEdges: src = rand.randrange(0, numVertices) dst = rand.randrange(0, numVertices) if src != dst: edges.add((src, dst)) return edges if __name__ == "__main__": """ Usage: transitive_closure [partitions] """ sc = SparkContext(appName="PythonTransitiveClosure") partitions = int(sys.argv[1]) if len(sys.argv) > 1 else 2 tc = sc.parallelize(generateGraph(), partitions).cache() # Linear transitive closure: each round grows paths by one edge, # by joining the graph's edges with the already-discovered paths. # e.g. join the path (y, z) from the TC with the edge (x, y) from # the graph to obtain the path (x, z). # Because join() joins on keys, the edges are stored in reversed order. edges = tc.map(lambda x_y: (x_y[1], x_y[0])) oldCount = 0 nextCount = tc.count() while True: oldCount = nextCount # Perform the join, obtaining an RDD of (y, (z, x)) pairs, # then project the result to obtain the new (x, z) paths. new_edges = tc.join(edges).map(lambda __a_b: (__a_b[1][1], __a_b[1][0])) tc = tc.union(new_edges).distinct().cache() nextCount = tc.count() if nextCount == oldCount: break print("TC has %i edges" % tc.count()) sc.stop()
I recently worked on a generator library for Elm (to simulate laziness and work with streams). Since I ended up publishing it as a package, I needed some examples and dug up the Sieve of Eratosthenes. ## Sieve of Eratosthenes The Ancient Greek algorithm provides a simple way to eliminate composite numbers (or sieve them out), leaving only the primes. Given the natural numbers > 1… • two is the first prime number; cross out multiples of two [2, 4, 6…] • the next number that is not crossed out – three – must be a prime; cross out multiples of three [3, 6, 9…] • etc ## Implementation Considerations How should the algorithm be implemented? In an imperative language with efficient side effects (e.g. array read writes), multiples can be crossed off all at once (though a finite upper bound must be specified). In purely functional contexts, a typical solution uses laziness with infinite recursion, e.g. from the Haskell Wiki: import Data.List.Ordered (minus, union, unionAll) primes = 2 : 3 : minus [5,7..] (unionAll [[p*p, p*p+2*p..] | p <- tail primes] In Elm 0.19, as far as I can tell, such infinitely recursive formulations are impossible with the strict compiler. Using generators, though, we can still implement a genuine, infinite sieve. ## Algorithm An algorithm for a purely functional, incremental sieve is described in section 3 of Melissa O’Neill’s paper, The Genuine Sieve of Eratosthenes. The idea is store the next known composite in a map. As larger candidates are explored and the known composites are found, the map is updated to reflect the next known composites (in a “just-in-time” manner). (A step-by-step illustration follows later in this post.) The Haskell implementation from the article for an initial version is relatively clear: sieve xs = sieve’ xs Map.empty where sieve’ [] table = [] sieve’ (x:xs) table = case Map.lookup x table of Nothing −> x : sieve’ xs (Map.insert (x*x) [x] table) Just facts −> sieve’ xs (foldl reinsert (Map.delete x table) facts) where reinsert table prime = Map.insertWith (++) (x+prime) [prime] table The Elm version incorporates a few optimizations noted in the article, including the fact that different “wheels” may be used to generate candidates. • trivially, start with 2 and only try odd numbers [3, 5, 7…], thereby excluding composites involving 2 (referred to as “wheel2”) • similarly, start with the first primes 2, 3, 5 and only try numbers that are not their composites [7, 11, 13…] (referred to as “wheel2357”) O’Neill’s paper provides a cycle that can be used to implement a wheel that excludes multiples of 2, 3, 5, and 7. The core of the Elm version is the recursive function that advances the prime number generator one step (thereby emitting the next prime and updating the map of composites). The logic is the same as the Haskell version above. Here, wheel is itself a generator, emitting candidates to be verified. sieveNext : SieveState b -> Maybe ( Int, SieveState b ) sieveNext ( map, wheel ) = let ( guess, wheel_ ) = in case Dict.get guess map of Nothing -> Just ( guess, ( insertNext guess wheel_ map, wheel_ ) ) Just composites -> sieveNext ( updateMap guess composites map, wheel_ ) Running elm repl on a modern dual-core laptop, the millionth prime can be found in under a minute. > import Generator as G > import Examples.Eratosthenes as E > E.wheel2357Init |> E.sieve |> G.take 10 [2,3,5,7,11,13,17,19,23,29] : List Int > E.wheel2357Init |> E.sieve |> G.drop 999999 |> G.take 1 [15485863] : List Int ## Observing the Algorithm A benefit of using a generator is the ease of observability of internal state at each step. Using the wheel of odd numbers only, the first two steps yields the first prime, 2, and the second prime, 3, at which point the composite map has been initialized. For the prime 3, the map is keyed at 3 * 3 (since smaller composites will have been previously checked), with a value being another generator for 3 * the wheel candidates, i.e. only the multiples of 3 that will be checked by the algorithm. > E.wheel2Init |> E.sieve |> G.advance 2 ( [2, 3] , Active { next = <function> , state = ( ( Dict.fromList [ (9, [Active { next = <function>, state = ((),3) }]) ] , Active { next = <function>, state = ((),3) } ) , [3] ) } ) Advancing a few steps, we see that the algorithm has moved past 9, the first composite in the map. Accordingly, the first key in the map has been updated to 15, to reflect the next multiple of 3 to be verified (3 * 5). Other squares of found primes (5, 7, 11) have also been added to the map, with their respective composite generators. > E.wheel2Init |> E.sieve |> G.advance 5 ( [2,3,5,7,11] , Active { next = <function> , state = ( ( Dict.fromList [ (15,[Active { next = <function>, state = ((),5) }]) , (25,[Active { next = <function>, state = ((),5) }]) , (49,[Active { next = <function>, state = ((),7) }]) , (121,[Active { next = <function>, state = ((),11) }]) ] , Active { next = <function>, state = ((),11) }) , [] ) } ) But we can do better. Since we’re already in Elm, why not feed the data directly into a visualization? (No need to bother with cumbersome serialization / deserialization to and from some non-native data structure.) ## Visualizing the Algorithm Since the goai is to facilitate the understanding of the sieve and composite map state at each step, the information is presented directly with tables, with highlighting to denote changes from the previous step. In the below screenshot, we see that with wheel2, 105 has multiple prime factors 3, 5, 7 – all of which have just been updated in the map to their next composites, highlighted in green.
Known as the marine isotope stage 11 (MIS 11), the interglacial period centered around 400,000 years ago was the longest and possibly the warmest interglacial in the past 0.5 million years. Because the orbital configurations, atmospheric greenhouse gas concentrations, climate, and faunal characteristics during MIS 11 closely resemble those of the past 5,000 years, paleoclimatologists use MIS 11 as a geological analogue of the present and the near future. There exist several high-resolution records documenting almost all aspects of terrestrial and marine climate through MIS 11. However, there is neither a clear understanding about how climactic parameters such as atmospheric carbon dioxide (CO2), sea surface temperature, the isotopic makeup of carbon in marine and terrestrial reservoirs, and annual air temperature interact, nor a consensus regarding the major drivers of climate change during this interval. Using 15 of the most robust proxy records of marine and terrestrial climate, Das Sharma et al. employ new statistical and mathematical techniques to quantify the interactions among climatic parameters and to investigate which of these parameters could be the primary drivers of climate change during MIS 11. The authors find that atmospheric CO2 concentration was indeed the primary driver of both terrestrial and marine climate: Sea surface temperature and the isotopic makeup of carbon in terrestrial and marine reservoirs responded "instantaneously" (i.e., within 1,000 years) to changes in atmospheric CO2 content. They further report that MIS 11 had warm and cool phases that can be detected from sea surface temperature records alone. During the relatively cold phases, sea surface and air temperatures behave coherently and respond to atmospheric CO2 faster. However, during warmer intervals, ocean surface and air temperatures behave more independently of each other and atmospheric CO2. The authors suggest that over the course of the next century, air and sea surface temperatures are likely to change in ways that will be difficult to predict. - S. Das Sharma, D. S. Ramesh, C. Bapanayya, P. A. Raju. Sea surface temperatures in cooler climate stages bear more similarity with atmospheric CO2forcing. Journal of Geophysical Research, 2012; 117 (D13) DOI: 10.1029/2012JD017725 Cite This Page:
"Joy to your soul” reads a note next to a newspaper clipping in a scrapbook belonging to Joy McLean Bosfield (1924-1999), a musician, educator, and entrepreneur who called Washington, DC home between 1962 and 1985. Joy is a consistent theme in two scrapbooks that she donated to the Anacostia Community Museum. Photos, plane tickets, programs, posters, and poetry chronicle a musical career that took her from New Jersey to Broadway, and then around the world in the 1940s and 1950s. Additional materials from later in her career include a long letter to curator Portia James. Her editorial expertise is evident in lyrical and sometimes funny explanations of how elements of her story connect. Explore Joy McLean Bosfield's scrapbooks, transcribed by Smithsonian digital volunteers. Look for her given name, Joy L. Mearimore, in concert programs from the 1940s, and her first stage name, Joy Merrimore, in the Broadway playbill for Memphis Bound!, in which she played Eulalia. She stars as Joy McLean, first in Great Britain in the late 1940s and early 1950s and then on tour with the international production of Porgy and Bessin the early to mid-1950s. Decades later in the District, she directs her next namesake, the McLean Bosfield Vocal Studios. Her annotations provide a guide to understanding her life's journey and insight into her skill as an educator and mentor. The Joy McLean Bosfield Papers also include phonograph records and archival materials from later in her career. In a long letter to curator Portia James accompanying her papers, McLean Bosfield's editorial expertise is evident in her lyrical explanation of how elements of her story connect.
# Define a procedure, add_to_index, # that takes 3 inputs: # - an index: [[<keyword>,[<url>,...]],...] # - a keyword: String # - a url: String # If the keyword is already # in the index, add the url # to the list of urls associated # with that keyword. # If the keyword is not in the index, # add an entry to the index: [keyword,[url]] ''' [ e0[<keyword>, [<url>,url2]], e1[keyword2, [url, url2]] e2[keyword3, [url, url2]] e3[keyword4, [url, url2]] ] ''' index = [] def add_to_index(index,keyword,url): appendedUrl = False for keyword_listUrls in index: # keyword already exists in index list if keyword == keyword_listUrls[0]: if url not in keyword_listUrls[1]: appendedUrl = True keyword_listUrls[1].append(url) # if we didn't find the keyword in the list, add it. if appendedUrl == False: index.append([keyword,[url]]) add_to_index(index,'udacity','http://udacity.com') add_to_index(index,'computing','http://acm.org') add_to_index(index,'udacity','http://npr.org') print index #>>> [['udacity', ['http://udacity.com', 'http://npr.org']], #>>> ['computing', ['http://acm.org']]]
On the left is a normal testicle. On the right is one that has been blasted with ultrasound. The latter has a reduced sperm count, which is why researchers at the University of North Carolina think that they may have discovered an effective male contraceptive: They found that two, 15-minute doses "significantly reduced" the number of sperm-producing cells and sperm levels. [...] Lead researcher Dr James Tsuruta said: "Further studies are required to determine how long the contraceptive effect lasts and if it is safe to use multiple times." The team needs to ensure that the ultrasound produces a reversible effect, contraception not sterilisation, as well as investigate whether there would be cumulative damage from repeated doses. Link -via Popular Science | Photo: James Tsuruta and Paul Dayton
- Breasts/Boobs/Bust/Mammaries Hypnotism - Mesmerizing Breasts/Boobs/Bust/Mammaries The user can put others in a suggestive trance upon visual contact with the users breasts, effect may take a few minutes to work or happen instantly. While in the trance, the person becomes highly pliant to their suggestions. - Mental Inducement - Mental Manipulation - Mesmerizing Presence - Mind Control - Pectukinetic Combat - Subliminal Seduction - Succubus Physiology - Supernatural Bodily Aspects - Users of Indomitable Will/Psychic Shield (highly resistant) and Psychic Immunity (impervious). - Depend on visual contact. - May only work on people of a certain gender or sexual orientation. - Users of Control Immunity may be immune. See Also: I Have Boobs You Must Obey. - The Mighty Endowed/Nina Dowd (DC Comics) - Meiko Ginya (Keijo!!!!!!!!) - Mio Kusakai (Keijo!!!!!!!!) - Okami (Manyuu Hikenchou) - Sarah (Misfits) - Areola (Stroker and Hoop)
है। 3) रात में एक चम्मच सौंफ आधा कप पानी में भिगो दें सुबह सौंफ को मसलकर पानी में छान लें अब इस पानी को दूध में मिलाकर पिलाने से बच्चे का पेट फूलना और गैस बनना ठीक हो जाता है आरा। बेवजह गोलियां छोड़ने वाले अनुज्ञप्तिधारियों के लाइसेंस की अनुज्ञप्ति बहुत जल्द रद होगी। पुलिस ऐसे अनुज्ञप्तिधारियों को चिह्नित कर शस्त्र अनुज्ञप्ति रद के लिए जिलाधिकारी के पास प्रस्ताव समर्पित करेगी। सोमवार को आयोजित क्राइम मीटिग में भोजपुर एसपी सुशील कुमार ने इसके साफ संकेत दिए है। उन्होंने सभी थानेदारों को आदेश दिया हैं कि ऐसे सभी अनुज्ञप्तिधारियों को चिह्नित कर प्रस्ताव समर्पित करें, जिन्होंने शस्त्रों के नियमों का दुरूपयोग किया है। उन्होंने बताया कि थानों में लाइसेंसी शस्त्रों के सत्यापन के दौरान पाया गया हैं कि अनगिनत ऐसे अनुज्ञप्तिधारी हैं जिन्होंने अपने लाइसेंस पर 500 से लेकर 700 तक गोलियों का उठाव किया है। वर्तमान में उनके पास पांच से दस गोलियां ही पाई गई हैं। ऐसे में शस्त्रों के नियमों का दुरूपयोग का मामला बनता है। एसपी ने बताया कि शस्त्रों की अनुज्ञप्ति आत्मरक्षार्थ के लिए दी जाती है। लेकिन, किसी ने आत्मरक्षा फायरिग नहीं की है। ऐसे में यह स्पष्ट रूप से पता चला रहा कि</s>
3600 मुफ्त कनेक्शन दिए जा चुके हैं। अब तक लगभग 4500 कनेक्शन दिए जा चुके हैं, हर शनिवार को यह कैंप लगाया जाता है, इसलिए अगर किसी को भी मुफ्त कनेक्शन लेना है तो वह अपने बिजलीघर के अवर अभियंता से सम्पर्क कर सकते हैं। एरोड्रम पुलिस के अनुसार 8 नवंबर13 को इलाके में रहने वाली 8वीं की छात्रा ने दोपहर 3 बजे फांसी लगाई थी। तात्कालिक जांचकर्ता एसआई गुर्जर ने बताया कि छात्रा घटना के दिन सुबह 10 बजे घर से कहीं गई थी और दोपहर करीब 3 बजे लौटी थी। तब उसकी मां ने पूछा था तो वह काफी गुस्से में थी और तीसरी मंजिल स्थित कमरे में चली गई थी। एक घंटे बाद पिता उससे बात करने पहुंचे तो दरवाजा बंद था और वह फंदे पर झूल थी। पुलिस ने पोस्टमार्टम करवाकर स्लाइड राऊ स्थित प्रयोगशाला में भेजी थी। रिपोर्ट में बलात्कार होना पता चला। अब जांच एसआई डामोर कर रही हैं। खासकर यह जगह हनीमून कपल्स के बीच खासा प्रसिद्ध है। इस सिटी को व्हाइट सिटी इसलिए कहा जाता है क्योंकि उदयपुर में ज्यादातर इमारतें संगेमरमर की बनी हैं इसलिए इसको यह नाम दिया गया है। वैसे इसको वेनिस ऑफ इंडिया भी कहा जाता है। युजिन</s>
What Every Parent Should Know About Education: How knowing the evidence can help your child succeed (Practical Teaching): How knowing the facts can help your child succeed£14.99* - A really handy book for parents who are trying to understand and grapple with decisions about which school may be best for their child. - The book is focused on the education system with England, helping parents understand the curriculum, expectations and assessment for each stage. - Questions are offered for parents to ask schools, leaders and teachers to ensure it is the best fit for their child. - The authors warn of debunked theories that are still advocated in some schools, offering advice on key attributes parents should listen out for. - The book is a digestible, supportive and accessible guide that can support parents from all backgrounds to understand what happens within the school system. What UKEdChat says about this book: When options are available, it’s always nice to research to help inform a positive decision. The same is now especially true in England’s education system, where many parents now have a range of choices which school to send their children to. In a system that is forever evolving, and the power of online reports, forums, and information, making the best choice for the education of the child can seem overwhelming. In their new book, Chris Atherton and Stuart Kime address some of these concerns, offering guidance, advice, and key questions that parents should ask of school if they are thinking about the choice available. Essentially exploring key areas, such as assessment and testing; learning; specific learning disabilities; behaviour, stress and mental health; and considerations between primary and secondary schools, the book brings together key information, experiences and key knowledge that can help towards an informed decision. The book is a digestible, supportive and accessible guide that can support parents from all backgrounds to understand what happens within the school system. The authors offer probing questions for parents that they should consider (and ask) when approaching schools, speaking to other parents, or come across information that may not be understood. Offering further signposts and references, this is a great book that should be considered by parents who feel overwhelmed with choosing the right school for their children. What the publisher says about this book: - What do you really know about how children learn? - How helpful are different types of assessment and what do the results mean? - Is homework necessary and how you can you encourage your child at home? - Will ability groups and setting help your child achieve more? - How do you choose the right school to suit your child? The government and the media have a lot to say about education, but what is the evidence behind these debates? This book walks you through all the most important issues in education, comparing commonly-held beliefs with simple summaries of the evidence, providing you with clear, jargon-free information. It covers topics including school choice, testing and assessment, homework and revision, primary- and secondary-specific topics, stress and mental health, and special needs. Most importantly it ensures you will be able to ask schools the right questions, interact positively with teachers and effectively support your child throughout their education.
Reimagining Modern Hebrew Instruction It’s no secret that though our students are exposed to years of modern Hebrew instruction, many emerge with relatively little comprehension of oral or written Ivrit to show for it. Too many can neither understand nor join an impromptu Israeli conversation, much less read, comprehend or write basic Hebrew. Many are doubtful about their prospects for gaining proficiency and are reluctant to use their lagging language skills in an authentic context. Of course, the study of Hebrew is just like the study of Spanish, Mandarin or any other language in this regard. A steady diet of vocabulary lists, grammar and rigorous testing rarely produces enthusiastic or proficient language listeners, readers, writers or speakers. Nor has dry cultural or religion-based content study in the modern Ivrit class contributed much to conversational competence. And yet we continue to rely upon these stock curricular ingredients, now with new-and-improved online games. Our unfulfilled mission frustrates Ivrit teachers as much as it hurts our students. We signed up to spread our passion for modern Hebrew, yet we never get to pluck the fruit of our toil. On the contrary, we often unwittingly contribute to negative attitudes around Ivrit and an overwhelming malaise: “It has been and always will be this way. Modern Hebrew just can’t be taught successfully.” Aligning with the Research What would it take to change this disheartening narrative and break the cycle of ineffective Hebrew offerings? How can we reimagine Ivrit instruction? The field of language teaching has evolved since the days of vocabulary lists and grammar drills. We are now poised to overhaul and revitalize the Hebrew quest by vastly improving teacher training and efficacy. A better-supported Ivrit faculty, grounded in knowledge of Second Language Acquisition (SLA), is the key to turning around our outmoded sequence and delivering the once elusive prize: transformed Hebrew attitudes and outcomes. While many Ivrit teachers may be unfamiliar with SLA, the good news is, it’s fairly intuitive. This patent precept anchors the work: We acquire language by understanding messages. Furthermore, the more compelling the messages, the more student attention to and engagement with them. With this framework in mind, an ever-expanding repository of research-aligned and classroom-tested instructional strategies is already successfully employed in language classrooms—including some Hebrew—around the world. The first phase of effective Hebrew professional re-development involves educating both institutional decision-makers and Ivrit instructors in SLA essentials. Only by internalizing this bread-and-butter of the profession can instructors identify which legacy elements and practices to scrap and how to build anew. Getting a Feel for It Once faculty and their supervisors understand the basic principles of SLA, teachers will next focus on honing their classroom communication skills. Delivering compelling and understandable oral messages is their number one task. One of the most practical ways to train faculty is by having master comprehension-based instructors model extended language lessons, but in a target language the teachers do not already speak. Experiencing initial uncertainty during a barrage of unfamiliar sounds and language chunks not only builds empathy for their students’ task, but trainees get an inspiring first-hand appreciation for the effectiveness of comprehensible-input-based strategies. This demonstration usually leaves enthusiastic teachers hungry to tinker in their own classrooms. Abandoning Unaligned Materials To this end, enlightened Ivrit teachers relinquish the prescriptive grammar and vocab-heavy textbook. Instead, they use narrow and repetitive extemporaneous language, in much the same way parents use “motherese” with their young children: concrete “here and now” messaging; a slowed pace; shortened/simplified sentences; and extralinguistic supports such as gestures, facial expressions, voice intonation, visuals, drama and movement. In short, the teacher-as-language-parent uses whatever means available to make Hebrew understood in real time. Such classes are conducted almost exclusively in Ivrit as our students, like babies, need to hear and simultaneously understand lots of Hebrew in order to begin acquiring it. Kids as Curriculum But how do we get our kids to tune in and sustain their attention while this extended oral language is coming in? Here is a real shift in how we conceive curriculum. Knowing that kids are most interested in talking about themselves—their experiences, ideas and opinions—we talk to and about them, extending the conversation to include everyone. Our kids are, especially in the beginning, the Ivrit curriculum! At first, teachers may build a bridge of interest by asking countless Hebrew questions—surveying, comparing and contrasting, voting, tallying and graphing results. Do you have a pet? What’s your favorite kind of candy? Cubs or Sox? Who’s your least favorite book character? Each inclusive poll invites students to share their personal response aloud, which the teacher massages back into extended conversation as she reports back to class. “Class, Esther says she loves Hermione Granger! Ya’akov says he likes Harry Potter, but Shira loves Hermione. Whom do you love, Ze’ev?” A single compelling query can yield cordial conversation, mostly provided by the teacher-as-language-model. This input is rife with narrow repetition, personal information and pleasant, community-building banter. Furthermore, such canvassing likely employs some of the highest frequency words, giving a real workout to practical sentence building-blocks, such as likes, has, says and wants. This levity and stress-free ambience support our SLA-informed goals of focusing primarily on meaning, not on language accuracy or mechanics. Reading Cart After Listening Horse But perhaps the most conspicuous shift away from traditional Hebrew instruction is delayed reading at the novice-to-intermediate levels. Gone are the days of densely stacked vocabulary words, glossed at the bottom of the page. Instead, our students wait to commence the literacy phase until they’ve interacted with and demonstrated comprehension of front-loaded oral language. Only then are students invited to read that which they have contributed to, and understood in class. With the teacher as reading guide, we ensure a successful and affirming reading experience that feels automatic and effortless. The specially prepared learner text or co-created story is often elicited from the group and written up by the teacher, who, having worked closely with her students, can accurately pitch readings to the class’s proficiency level. So, in our reimagined Ivrit classes, our kids turn off their devices and clear their desks; listen to, understand and interact with the compelling Hebrew input surrounding them, supported by visual anchors, such as pictures, props and student dramatization; and then read what they’ve been hearing, once we’re sure they confidently recognize both sound and meaning. A broad selection of aligned literacy extensions can provide even more novel repetition of this narrow language, affording students time to acquire, without hastily proceeding to the next unit of study. Creating a Chill Environment With the newly established priorities of student interest and ease of understanding in place, retooled Ivrit teachers debunk conventional wisdom that a sense of discernable struggle or rigor is conducive to acquiring Hebrew. On the contrary, we want our kids to understand Ivrit unconsciously and without frustration, and to that end we frequently check to insure they “get it.” This constant monitoring helps us adjust both our rate of delivery and register of speech, ensuring that no passengers fall from our Understanding Train. Learning a new language is a distinct process, different from, say, algebra or history. We want to safeguard the Ivrit class as an accessible, pleasant and seemingly easy environment, as the brain works hard to process lots of incoming Hebrew data. Unfortunately, though schools may attempt to improve the quality of their offerings by piloting new Ivrit curricula and attending materials-training sessions, this rehashed content often perpetuates the same pedagogically indefensible missteps of the past. Devoid of SLA undergirding, it tends to favor forced student output, writing and speaking, rather than prioritize listening and reading, the language inputs that drive acquisition. Stakeholders must, therefore, be grounded in SLA in order to recalibrate and embrace reasonable expectations for language production, which is the result of copious comprehended input. Once this shift in classroom focus and revised proficiency timeline is adopted, we will no longer assess beginner-to-intermediate student progress on how well or how much the kids can say or write, but rather on how well they understand increasingly sophisticated chunks of oral and written language at the discourse level. The community will simultaneously notice increased engagement, improved attitudes and a more pleasant classroom atmosphere with better teacher-student relationships. Eventually, natural unforced Hebrew output will emerge, in the service of friendly communication. Supporting the Chalutzim A paradigm shift of this magnitude can be daunting, and routine challenges are to be expected. A spirit of collaborative troubleshooting, in which Hebrew teachers have a voice in say, scheduling, teaching assignments, classroom configuration and technology setup can go far in assuaging underlying tensions even before new teaching strategies are brought to students. Once teachers are ready to embrace their reimagined roles, the sudden spotlight on a discipline that heretofore flew under the radar can be jarring. Planning for peer-collaboration time, support and observation plus regular team meetings with the supervisor and/or evaluating administrator can help bolster a growth mindset (Carol Dweck), problem-solve and identify areas for additional professional training. The first steps of an unfamiliar journey can feel vulnerable, so pioneer teachers should be encouraged and acknowledged. With this course change, Hebrew teachers are trying on new interpersonal behaviors, content, assessment and classroom management, and learning as they go. Regular and ongoing supervisory observation, feedback, coaching and mentoring is crucial, though teachers can also initiate cycles of powerful self-reflection by videorecording themselves teaching, and watching themselves later, with a qualitative rubric in hand. The key is for them to notice not only what they are doing as the leader in the room but how the students are responding and interacting with the Ivrit surrounding them. Communities of comprehension-based practice and texts are also widely available to help support our Hebrew teacher-innovators. Workshops and conferences targeting comprehension-based strategies for instruction abound, and the blogosphere teems with teacher articles, curricular resources and demonstration videos, including the first community blog specific to teaching comprehension-based Hebrew: cmovan.edublogs.org (article author’s blog). Because Hebrew is mostly taught in private settings, this grassroots Comprehensible Input movement is relatively new for Hebrew teachers, who, until now, have not trained alongside public school language teachers. Participation in this wider professional setting is a healthy venture that builds networks of support and inquiry and can also provide laboratories for additional peer observation and reflection. A Positive and Proactive School Culture is the Bedrock Finally, but no less significant, is the school’s institutional culture. Not only do we need to ask hard questions about the community’s apathy or negative attitude towards Hebrew instruction (and how to repair it through research-aligned teaching and learning). Educational and administrative leaders must also be sufficiently grounded in the new language instruction rationale to discuss and defend it within the wider community. By exemplifying engaged learning and unwavering commitment to improvement, leaders hold their principals, teaching supervisors, department chairs and faculty accountable to aligned practices. Realizing this vision of excellence often requires attendance at a professional conference or workshop alongside Hebrew faculty; hosting informational opportunities for parents to ask questions and have their concerns heard and answered; regular visits to classrooms and department meetings; frequent informal observations with qualitative feedback, perhaps initially modeled by an outside trainer; and defined and attainable goals for teacher progress along a continuum of improvement. Additionally, stakeholders must also reflect on the tone and tenor of relationships in school. By acknowledging that we are all shaped by the teacher, administrator, student and parent dynamics, school administrators at every level can do much to ensure that classrooms, hallways, meetings and assemblies reflect both the school’s academic mission and highest aspirations for community conduct. Rooted in a healthy and positive school culture, with determination and a renewed sense of purpose, we are set to transform our Hebrew dreams into reality. As Eliezer Ben Yehuda predicted, “The Hebrew language will go from the synagogue to the house of study… to the school… [I]t will come into the home and... become a living language.” A part of this article was originally printed in The Chicago Jewish Home.
# Given we observe $x$ successes in sample of $n$, what is the probability that $p>0.5$ I flip a coin $$n$$ times, and I find that for $$x$$ of these it lands on heads. Let $$p$$ be the true probability of the coin landing on heads, though this is currently unknown. I am interested in $$P(p>0.5 \mid$$ I flipped it $$n$$ times seeing $$x$$ occurrences of heads$$)$$. I would really appreciate if someone could help me with: (a) Working this probability out (b) If there's a function in R I could use to do this (inputting the $$n$$ and $$x$$ I have for the different coins). • You need to supply at least one more piece of information: what do you currently think $p$ is and how strongly do you believe that? This information is usually encoded in a prior distribution. If you don't have such a distribution in mind, then you are probably asking the wrong question: instead, you might be looking for a hypothesis test. – whuber May 20, 2020 at 15:09 • Thanks for the quick reply! I didn't have an expected value of $p$ in mind, all I'm interested in is given I've seen $x$ out of $n$ successes, how likely it is that the value of $p$ is greater than $0.5$. I'm not sure I understand why I'm not able to work out such a probability from the information I provided. For example if I observed $n = x = 100$, I'd expect the probability that the 'true value of $p$ is larger than $0.5$' to be fairly high (though I have no idea how to actually calculate it). May 20, 2020 at 15:23 • Well, you have observed $p'=\frac{x}{n}$ successes. If $p'>p$, your probability is 1, else it is 0. Either way, you are done. Does this answer your question? If not, your counter-argument is the beginning of the discussion of why you need a prior probability. May 20, 2020 at 15:27 • Sounds like I need to read up about prior probabilities some more! For now though, given I have no further context around whether this coin is biased, the sensible thing to do seems to be to take the prior distribution to be binomial with $p=0.5$. I'll update the question to reflect this! May 20, 2020 at 15:42 Well the probability that you are asking for is not computable, at least not using frequentist statistics. Observe that in general, a conditional probability is computed as $$P(A|B)=\dfrac{P(A\cap B)}{P(B)}$$ But here, $$P(B)=$$ I flipped the coin n times seeing x occurrences of heads. And for this probability we would need to know the value of $$p$$. You can, as sugested in the comments, consider a Bayesian perspective, but for doing that you would require knowledge on a prior distribution. What you can actually do from a frequentist perspective is to comupte a hypothesis test. For doing this, you should define your null hypothesis: $$H_0: p=0.5$$ And the alternative hypothesis , which is: $$H_1:p>0.5$$ Now we can compute the test statistic $$T = \frac{\hat{p}-p_{0}}{\sqrt{p_{0}\left(1-p_{0}\right) / n}} \sim_{a p} N(0,1)$$ where $$p_0=0.5$$ (the value under your null hypothesis) and $$\hat{p}=x/n$$ is the sample proportion. plugging in the values for $$p_0$$ annd $$\hat{p}$$, T is an actual number. Now consider a significance level $$\alpha = 0.05$$ for your test. Then you can check using for example, the program R, that $$z_{\alpha}=1.64$$. This means that $$P(N(0,1)>1.64) = 0.05$$ Finally, assuming that the null hypothesis is true, T is drawn approximately from a normal distribution, so you can compare T against the value you would expect from a normal distribution, which is 1.64 • If $$T>1.64$$ then T is a strange value, because it has a probability of being from a normal distribution smaller than 5%. You reached this conclusion assuming that the null hypothesis was true, so you conclude that the null hypothesis should be rejected and the alternative hypothesis should be supported. • If $$<=1.64$$ you cannot reject the null hypothesis. Doing this in R: n = 100 # number of tosses x = 60 # number of heads p = 0.5 # your null hypothesis probability binom.test(x=x, n=n, p=p, alternative='greater', conf.level=0.95) Which returns Exact binomial test data: x and n number of successes = 60, number of trials = 100, p-value = 0.02844 alternative hypothesis: true probability of success is greater than 0.5 95 percent confidence interval: 0.5129758 1.0000000 sample estimates: probability of success 0.6 So the p-value here is $$0.02844$$. If the p-value is smaller than $$\alpha=0.05$$ then you do not reject the null hypotheis. If the p-value is greater than $$\alpha$$ then you reject the null hypothesis
This study examines the impacts arising from neuroscience and mental health research going back 20–25 years, and identifies attributes of the research, researchers or research setting that are associated with translation into patient benefit, in the particular case of schizophrenia. This report presents the full set of backward-tracing perspectives. The study combined two methods: forward-tracing case studies to examine where scientific advances of 20 years ago have led to impact today; and backward-tracing perspectives to identify the research antecedents of today's interventions in schizophrenia. These research and impact trails are followed principally in Canada, the UK and the USA. The headline findings are as follows: - The case studies and perspectives support the view that mental health research has led to a diverse and beneficial range of academic, health, social and economic impacts over the 20 years since the research was undertaken. - Clinical research has had a larger impact on patient care than basic research has over the 20 years since the research was undertaken. - Those involved in mental health research who work across boundaries are associated with wider health and social benefits. - Committed individuals, motivated by patient need, who effectively champion research agendas and/or translation into practice are key in driving the development and implementation of interventions. The project described in this report was supported in Canada by the Graham Boeckh Foundation, Alberta Innovates Health Solutions, and the Canadian Institutes of Health Research; in the UK by the National Institute for Health Research; and in the USA by the National Institute of Mental Health. The research was conducted by RAND Europe. This report is part of the RAND Corporation research report series. RAND reports present research findings and objective analysis that address the challenges facing the public and private sectors. All RAND reports undergo rigorous peer review to ensure high standards for research quality and objectivity. Permission is given to duplicate this electronic document for personal use only, as long as it is unaltered and complete. Copies may not be duplicated for commercial purposes. Unauthorized posting of RAND PDFs to a non-RAND Web site is prohibited. RAND PDFs are protected under copyright law. For information on reprint and linking permissions, please visit the RAND Permissions page. The RAND Corporation is a nonprofit institution that helps improve policy and decisionmaking through research and analysis. RAND's publications do not necessarily reflect the opinions of its research clients and sponsors.
class RingBuffer: def __init__(self, capacity): self.storage = [] # list for holding items self.capacity = capacity # max number of items the buffer can hold self.curr_index = 0 # attribute for keeping track of where to add items when capacity is reached # Method for adding items to a Ring Buffer def append(self, item): if len(self.storage) < self.capacity: # if the number of items in the buffer are less than the capacity then add the item to the end self.storage.append(item) else: # if capacity is reached then overwrite values in the buffer with newly added items self.storage[self.curr_index] = item # calculate the new current index, example capacity = 4 index = 3 + 1, 4 % 4 = 0, so index goes back to start self.curr_index = (self.curr_index + 1) % self.capacity # Method for retrieving all the items in a RingBuffer and returning them in a list def get(self): # Create array for holding items in buffer temp = [] for value in self.storage: # for every value in the buffer then add it to temp if it is not None if value: temp.append(value) # return a list of the values currently in the buffer return temp rb = RingBuffer(5) # Initialize a Ringbuffer with capacity set at 5 rb.append(7) # 7 rb.append(14) # 7 14 rb.append(21) # 7 14 21 rb.append(28) # 7 14 21 28 rb.append(35) # 7 14 21 28 35 (Capacity met) rb.append(42) # 42 14 21 28 35 rb.append(49) # 42 49 21 28 35 rb.append(56) # 42 49 56 28 35 print(rb.get()) # Should print: [42, 49, 56, 28, 35]
WHEN IS THE MEDIAN NOT A GOOD MEASURE OF CENTRAL TENDENCY? The median can be a good measure of central tendency, especially when the data is evenly distributed. It can illustrate a skewed and uneven distribution when compared to the mean. However, there are occasions when the median is not always a good measure of central tendency. As the median only reports on the middle most observation it does ignore the actual values of all the rest of the data, one outcome is that the median may hide the presence of extreme values. Let us consider the wages of 'The Widget Company', we will increase the earnings of the CEO from $100,000 to $500,000. When the CEO's salary changes to $500,000 we can see that the median income remains at $50,000, even though the CEO now earns five times their previous income.
#!/usr/bin/env python3 """Normalize contants of a matrix""" import numpy as np def normalization_constants(X): """Normalize constants of a matrix - X is the numpy.ndarray of shape (m, nx) to normalize - m is the number of data points - nx is the number of features Returns: the mean and standard deviation of each feature, respectively """ return np.mean(X, axis=0), np.std(X, axis=0)
I may have already posted something similar to this last yr. But anyway....LOVED this movie! And if you know anything about ancient Sumer and the Mesopotamian culture it added even more to the storyline. Remember, the alien did state: "I am Anu....I AM............God." He was a chief sky god. I think it is a multifaceted movie and was meant to be interpreted in many ways, everything symbolic had more than one meaning. Yes, they large eyes do remind many of the grays. But also the owl played a large part in ancient culture. It was often seen as an ill omen (I'm not referring to Lilith). These are descriptions of foul spirits (no pun intended)taken from ancient Mesopotamian: "They are gloomy, their shadow dark, no light is in their bodies, ever they slink along covertly, walk not upright, from their claws drips bitter gall, their footprints are (full of) evil venom.. Neither males are they, nor females, they are winds ever sweeping along, they have not wives, engender not children, know not how to show mercy, hear not prayer and supplication.... The shivers and chills (of death) that fritter the sum of things, spawn of the god of heaven, spawned on an evil spirit, the death warrants, beloved sons of the storm god, born of the queen of the netherworld, who were torn out of heaven and hurled from the earth as castoffs, are creatures of hell, all. Up above they roar, down below they cheep, they are the bitter venom of the gods, they are the great storms let loose from heaven, they are the owl(of ill omen)that hoots in the town, spawn spawned by the god of heaven, sons born by earth are they. Over high roofs, over broad roofs like a flood wave they surge, from house to house they climb over, Doors do not hold them, locks do not restrain them... through the doors they glide like snakes, through the hinge boxes they blow like wind. From the man's embrace they lead off the wife, from the man's knee they make the child get up, and the youth they fetch out of the house of his in-laws, they are the numbness, the daze, that tread on the heels of man." They have been here for a very long time.
Your trading plan is THE critical system that you must work on and has the aim of guiding trader action in the heat of market decision making. Errors in trading plan content and construction can impact on consistency and the ability to ascertain what plan components are working for you or otherwise.Join Mike Smith of GO Markets for this innovative session that aims to give you: - A review of the key purpose and function of trading plans and why they are crucial to traders of all instruments. - Absolute clarity about what these 5 errors are and the potential impact on results if not addressed. - Practical trading plan statement examples of how to take your trading plan to the next level.
Student experience has a central focus in higher education as institutions strive for retention, attainment, and growth; acknowledging students as both consumers and participants (Muijs et al. 2017). Consequently, universities pursue active engagement of student voice to inform quality enhancement as a route toward improving educational experiences. Within the Scottish sector, quality enhancement is defined in part as the deliberate steps taken “to bring about improvement in the effectiveness of the learning experiences of students” (QAA, 2017, p. 3). In the School of Education and Social Work at the University of Dundee, an interdisciplinary team has initiated a three year study, exploring the experiences of a previously overlooked student group (Morgan et al. 2014), the taught postgraduate (PGT) student population, consisting of in-service professionals including early and middle career teachers pursuing a Master’s degree. This study aligns with the University vision, values and core principles, aiming to enhance student personal development, well-being and performance, by exploring student experiences to inform quality enhancement. To this end, a participatory action research design has been adopted to engage and actualize student voice. Using an online survey followed by focus groups, we have begun to explore the phenomenology of student experience across five areas emerging from a review of literature (Morgan, 2014; QAA, 2017; Temple et al., 2014): Transitions, Academics, University Services, Culture and Daily Life. To further explore student experience, well-being measures have been collected, reflecting the understanding that well-being is important in human functioning. According to Boekharts (1993) negative emotions arise when an experience is perceived as a threat to an individuals’ well-being. When this happens, the negative impact on well-being becomes the focus of attention, disrupting capacity for learning. Through this understanding, we posit a relationship between student experience, well-being, and engagement with learning. Findings from the first year of the study highlight the formative nature of PGT education; they indicate cognitive and affective processes underpin the student journey directly influencing student experience through three distinct dimensions: Degree of Connectedness; Effectiveness of Supports; and Quality of Communications. Resoundingly, students report feelings of transformation through their student journey. Transformation in thinking, understanding, capacity to problem solve and quality of social interactions. Simultaneously, students describe a range of heightened emotional peaks and troughs throughout their journey that at times act to heighten or threaten motivation and engagement. Through our analysis we sought to understand more clearly the challenges and threats students encounter and the various factors influencing their capacity to overcome. Three distinct dimensions of student experience emerged, each containing sub-categories operating independently and at times interacting to heighten or threaten the student journey. Three sub-categories define the degree of connectedness dimension. These include the students’ perceived sense of belonging, identity and relationships. Through these sub-categories students establish membership, academic identity, and comradery promoting resilience during high stress periods. Perceived availability of support, access to support and quality of support provided through University Services and within the programme of study define the effectiveness of supports dimension. Findings suggest a relationship between students’ perceived effectiveness of support and their capacity for self-directed learning. Effective supports, those defined as useful in promoting goal achievement, prompt increased motivation to engage in the study journey, while ineffective supports trigger feelings of demotivation, frustration and anxiety, sometimes lasting for extended periods of time, becoming impediments to their study journey. The third dimension, Quality of Communication, includes communication between student and University Service or within the student’s programme of study including tutor email exchanges, tutorials, and feedback. It is defined by students’ perceived sense of efficiency of communication, tone of communication, and clarity of response toward meeting the students’ needs. This dimension interacts heavily with the other dimensions through its capacity to impact academic identity, sense of membership, and self-efficacy. Acting as a threat, this dimension prompted feelings of confusion, demoralization, and isolation, placing the student journey at risk through decreased sense of self-efficacy, belonging and academic identity. Conversely, quality communication influenced student empowerment through increased confidence, and capacity to pursue self-directed learning. Interestingly, positive peer and tutor relationships emerged as the greatest factor promoting resilience in times of threat; through a foundation of strong peer and tutor relationships students report an improved capacity to persevere through high stress, overcoming threats from within any one of the dimensions, to continue on their study journey. Implications are twofold. For teacher educators, a revised understanding of the diverse nature and needs of this population, overcoming the taken-for-granted assumption that this group begins study possessing sound academic identities and the capacity to engage in self-directed learning, and that these qualities are inherently impregnable. Herein lies the potential for innovative programming, focusing on the relationship between student experience, well-being and learning. For the in-service teachers engaging in programmes designed through this revised understanding, resides the potential their learning experiences will be sustained through embedded supports fostering resilience throughout the student journey, leading to the formation of strong academic-identities, high self-efficacy, and sustained motivation to pursue professional learning. Boekaerts, M. (1993) ‘Being concerned with well-being and with learning’. Educational psychologist, 28(2), pp.149-167. Education Scotland (2017) Enhancement led Review Handbook (ELIR-4) (PDF) (Accessed: 10 September 2018). Quality Assurance Agency. Morgan, M. and Rigby, S. (2014) The importance of understanding the expectations and attitudes of the student body, university staff and business and industry in improving the STEM postgraduate taught student experience. (Accessed: 25 August 2018) The Higher Education Academy STEM. Muijs, D. and Bokhove, C. (2017) ‘Postgraduate student satisfaction: A multilevel analysis of PTES data’. British Educational Research Journal, 43(5), pp.904-930. Temple, P., Callender, C., Grove, L. and Kersh, N. (2014) Managing the student experience in a shifting higher education landscape. (PDF) (Accessed: 28 September, 2018) The Higher Education Academy.
Our goal in this paper is to present a generalization of the spectral zeta regularization for general Feynman amplitudes on Riemannian manifolds. Our method uses complex powers of elliptic operators but involves several complex parameters in the spirit of analytic renormalization by Speer, to build mathematical foundations for the renormalization of perturbative interacting quantum field theories. Our main result shows that spectrally regularized Feynman amplitudes admit analytic continuation as meromorphic germs with linear poles in the sense of the works of Guo–Paycha and the second author. We also give an explicit determination of the affine hyperplanes supporting the poles. Our proof relies on suitable resolution of singularities of products of heat kernels to make them smooth. As an application of the analytic continuation result, we use a universal projection from meromorphic germs with linear poles on holomorphic germs to construct renormalization maps which subtract singularities of Feynman amplitudes of Euclidean fields. Our renormalization maps are shown to satisfy consistency conditions previously introduced in the work of Nikolov–Todorov–Stora in the case of flat space-times. Cite this article Nguyen Viet Dang, Bin Zhang, Renormalization of Feynman amplitudes on manifolds by spectral zeta regularization and blow-ups. J. Eur. Math. Soc. 23 (2021), no. 2, pp. 503–556DOI 10.4171/JEMS/1016
Kengoro, the Robot That Sweats A robot named Kengoro was built in the University of Tokyo that was powered by more than 100 individual motors that made it flexible in terms of its movements. In its ability to produce sweat, researchers derived the idea when motors are active for long periods of time, they start to overheat. Eventually, robots require some sort of cooling system.Kengoro's frame was built out of laser-sintered aluminum powder which gave the researchers the ability to precisely control the structure of the skeleton. In order for the aluminum powder to be porous, researchers used lasers to fuse it into a solid. Since cooling systems are bulky which will limit the construction of lightweight movable robots, their idea for porous aluminum powder became their design for novel cooling system.Kengoro can work out for 11 minutes doing pushups without the need to stop and this is pretty good for a humanoid machineWant to know more about this awesome device? Follow this link! -Fun facts and Trivia brought to you by Research and Development Committee-
Sunday School for ages 3 through adults is held each Sunday at 10:15 am during the school year. 3 yr. olds through 5th grade learn Bible stories which relate to the preaching text for each week. Middle schoolers have a three-year rotation learning the Old Testament, New Testament, Life of Martin Luther and about prayer and worship. High Schoolers have a faith discussion format, and adults do a variety of Bible studies. Click HERE for the Sunday school schedule of events. Grace Kids (This program has not yet resumed due to Covid-19) GRACE KIDS is our new youth group for kids in Kindergarten through 5th Grade. All Kindergarten through 5th Graders are welcome to join us as we use TeamKID: All In to help grow young disciples by strengthening their relationships with God and with one another through education and having fun. It features Bible exploration through stories, games and fun activity-based learning all while helping kids grow to be more like Jesus. Grace Kids meets from 6:30-7:45pm on select Wednesdays throughout the year. Splash Ministry is a monthly mailing sent to families when a child is born. The Splash packets continue each month until children begin Sunday school and provide ideas for parents to begin faith learning with even these youngest of God’s beloved children. Milestone Ministry recognizes faith milestones in the life of Grace Lutheran children and youth. Students are recognized for their learning by receiving a stone. Milestones include: - Baptism and Remembrance of Baptism - 3 year-old receive story Bibles to begin to learn and love God’s Word - 4 year olds receive prayer pillows to help them learn to pray - Kindergarteners learn the Lord’s Prayer - 1st and 2nd Graders learn the 10 Commandments - 3rd Graders receive a Bible - 4th Graders learn about what God did for them in Baptism - 5th Graders learn about the Sacrament of Holy Communion - Camp and youth gathering milestones for those children and youth are grow their faith during the summer - Middle Schoolers and High Schoolers learn about leadership in the church by leading worship once a year - 10th Graders are confirmed - 12th Graders are recognized on their graduation
# -*- test-case-name: twistedcaldav.test.test_icalendar -*- ## # Copyright (c) 2005-2014 Apple Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ## """ iCalendar Utilities """ __all__ = [ "InvalidICalendarDataError", "iCalendarProductID", "allowedComponents", "Property", "Component", "tzexpand", ] import codecs from difflib import unified_diff import heapq import itertools import uuid from twisted.internet.defer import inlineCallbacks, returnValue from twext.python.log import Logger from txweb2.stream import IStream from txweb2.dav.util import allDataFromStream from twistedcaldav.accounting import accountingEnabledForCategory, \ emitAccounting from twistedcaldav.config import config from twistedcaldav.dateops import timeRangesOverlap, normalizeForIndex, differenceDateTime, \ normalizeForExpand from twistedcaldav.instance import InstanceList, InvalidOverriddenInstanceError from twistedcaldav.timezones import hasTZ, TimezoneException from txdav.caldav.datastore.scheduling.utils import normalizeCUAddr from pycalendar.icalendar import definitions from pycalendar.parameter import Parameter from pycalendar.icalendar.calendar import Calendar from pycalendar.icalendar.component import Component as PyComponent from pycalendar.componentbase import ComponentBase from pycalendar.datetime import DateTime from pycalendar.duration import Duration from pycalendar.exceptions import ErrorBase from pycalendar.period import Period from pycalendar.icalendar.property import Property as PyProperty from pycalendar.timezone import Timezone from pycalendar.utcoffsetvalue import UTCOffsetValue log = Logger() iCalendarProductID = "-//CALENDARSERVER.ORG//NONSGML Version 1//EN" allowedStoreComponents = ("VEVENT", "VTODO", "VPOLL",) allowedSchedulingComponents = allowedStoreComponents + ("VFREEBUSY",) allowedComponents = allowedSchedulingComponents + ("VTIMEZONE",) def _updateAllowedComponents(allowed): global allowedStoreComponents, allowedSchedulingComponents, allowedComponents allowedStoreComponents = allowed allowedSchedulingComponents = allowedStoreComponents + ("VFREEBUSY",) allowedComponents = allowedSchedulingComponents + ("VTIMEZONE",) # Additional per-user data components - see datafilters.peruserdata.py for details PERUSER_COMPONENT = "X-CALENDARSERVER-PERUSER" PERUSER_UID = "X-CALENDARSERVER-PERUSER-UID" PERINSTANCE_COMPONENT = "X-CALENDARSERVER-PERINSTANCE" # 2445 default values and parameters # Structure: propname: (<default value>, <parameter defaults dict>) normalizeProps = { "CALSCALE": ("GREGORIAN", {"VALUE": "TEXT"}), "METHOD": (None, {"VALUE": "TEXT"}), "PRODID": (None, {"VALUE": "TEXT"}), "VERSION": (None, {"VALUE": "TEXT"}), "ATTACH": (None, {"VALUE": "URI"}), "CATEGORIES": (None, {"VALUE": "TEXT"}), "CLASS": (None, {"VALUE": "TEXT"}), "COMMENT": (None, {"VALUE": "TEXT"}), "DESCRIPTION": (None, {"VALUE": "TEXT"}), "GEO": (None, {"VALUE": "FLOAT"}), "LOCATION": (None, {"VALUE": "TEXT"}), "PERCENT-COMPLETE": (None, {"VALUE": "INTEGER"}), "PRIORITY": (0, {"VALUE": "INTEGER"}), "RESOURCES": (None, {"VALUE": "TEXT"}), "STATUS": (None, {"VALUE": "TEXT"}), "SUMMARY": (None, {"VALUE": "TEXT"}), "COMPLETED": (None, {"VALUE": "DATE-TIME"}), "DTEND": (None, {"VALUE": "DATE-TIME"}), "DUE": (None, {"VALUE": "DATE-TIME"}), "DTSTART": (None, {"VALUE": "DATE-TIME"}), "DURATION": (None, {"VALUE": "DURATION"}), "FREEBUSY": (None, {"VALUE": "PERIOD"}), "TRANSP": ("OPAQUE", {"VALUE": "TEXT"}), "TZID": (None, {"VALUE": "TEXT"}), "TZNAME": (None, {"VALUE": "TEXT"}), "TZOFFSETFROM": (None, {"VALUE": "UTC-OFFSET"}), "TZOFFSETTO": (None, {"VALUE": "UTC-OFFSET"}), "TZURL": (None, {"VALUE": "URI"}), "ATTENDEE": (None, { "VALUE": "CAL-ADDRESS", "CUTYPE": "INDIVIDUAL", "ROLE": "REQ-PARTICIPANT", "PARTSTAT": "NEEDS-ACTION", "RSVP": "FALSE", "SCHEDULE-AGENT": "SERVER", }), "CONTACT": (None, {"VALUE": "TEXT"}), "ORGANIZER": (None, {"VALUE": "CAL-ADDRESS"}), "RECURRENCE-ID": (None, {"VALUE": "DATE-TIME"}), "RELATED-TO": (None, {"VALUE": "TEXT"}), "URL": (None, {"VALUE": "URI"}), "UID": (None, {"VALUE": "TEXT"}), "EXDATE": (None, {"VALUE": "DATE-TIME"}), "EXRULE": (None, {"VALUE": "RECUR"}), "RDATE": (None, {"VALUE": "DATE-TIME"}), "RRULE": (None, {"VALUE": "RECUR"}), "ACTION": (None, {"VALUE": "TEXT"}), "REPEAT": (0, {"VALUE": "INTEGER"}), "TRIGGER": (None, {"VALUE": "DURATION"}), "CREATED": (None, {"VALUE": "DATE-TIME"}), "DTSTAMP": (None, {"VALUE": "DATE-TIME"}), "LAST-MODIFIED": (None, {"VALUE": "DATE-TIME"}), "SEQUENCE": (0, {"VALUE": "INTEGER"}), "REQUEST-STATUS": (None, {"VALUE": "TEXT"}), "VOTER": (None, { "VALUE": "CAL-ADDRESS", "CUTYPE": "INDIVIDUAL", "ROLE": "REQ-PARTICIPANT", "RSVP": "FALSE", "SCHEDULE-AGENT": "SERVER", }), } # transformations to apply to property values normalizePropsValue = { "ATTENDEE": normalizeCUAddr, "ORGANIZER": normalizeCUAddr, "VOTER": normalizeCUAddr, } ignoredComponents = ("VTIMEZONE", PERUSER_COMPONENT,) # Used for min/max time-range query limits minDateTime = DateTime(1900, 1, 1, 0, 0, 0, tzid=Timezone(utc=True)) maxDateTime = DateTime(2100, 1, 1, 0, 0, 0, tzid=Timezone(utc=True)) class InvalidICalendarDataError(ValueError): pass class Property (object): """ iCalendar Property """ def __init__(self, name, value, params={}, parent=None, **kwargs): """ @param name: the property's name @param value: the property's value @param params: a dictionary of parameters, where keys are parameter names and values are (possibly empty) lists of parameter values. """ if name is None: assert value is None assert params is None pyobj = kwargs["pycalendar"] if not isinstance(pyobj, PyProperty): raise TypeError("Not a Property: {0!r}".format(property,)) self._pycalendar = pyobj else: # Convert params dictionary to list of lists format used by pycalendar valuetype = kwargs.get("valuetype") self._pycalendar = PyProperty(name, value, valuetype=valuetype) for attrname, attrvalue in params.items(): self._pycalendar.addParameter(Parameter(attrname, attrvalue)) self._parent = parent def __str__(self): return str(self._pycalendar) def __repr__(self): return ( "<{self.__class__.__name__}: {name!r}: {value!r}>" .format(self=self, name=self.name(), value=self.value()) ) def __hash__(self): return hash(str(self)) def __ne__(self, other): return not self.__eq__(other) def __eq__(self, other): if not isinstance(other, Property): return False return self._pycalendar == other._pycalendar def __gt__(self, other): return not (self.__eq__(other) or self.__lt__(other)) def __lt__(self, other): my_name = self.name() other_name = other.name() if my_name < other_name: return True if my_name > other_name: return False return self.value() < other.value() def __ge__(self, other): return self.__eq__(other) or self.__gt__(other) def __le__(self, other): return self.__eq__(other) or self.__lt__(other) def duplicate(self): """ Duplicate this object and all its contents. @return: the duplicated calendar. """ # FIXME: does the parent need to be set in this case? return Property(None, None, None, pycalendar=self._pycalendar.duplicate()) def name(self): return self._pycalendar.getName() def value(self): return self._pycalendar.getValue().getValue() def strvalue(self): return str(self._pycalendar.getValue()) def _markAsDirty(self): parent = getattr(self, "_parent", None) if parent is not None: parent._markAsDirty() def setValue(self, value): self._pycalendar.setValue(value) self._markAsDirty() def parameterNames(self): """ Returns a set containing parameter names for this property. """ result = set() for pyattrlist in self._pycalendar.getParameters().values(): for pyattr in pyattrlist: result.add(pyattr.getName()) return result def parameterValue(self, name, default=None): """ Returns a single value for the given parameter. """ try: return self._pycalendar.getParameterValue(name) except KeyError: return default def parameterValues(self, name, default=None): """ Returns a multi-value C{list} for the given parameter. """ try: return self._pycalendar.getParameterValues(name) except KeyError: return default def hasParameter(self, paramname): return self._pycalendar.hasParameter(paramname) def setParameter(self, paramname, paramvalue): self._pycalendar.replaceParameter(Parameter(paramname, paramvalue)) self._markAsDirty() def removeParameter(self, paramname): self._pycalendar.removeParameters(paramname) self._markAsDirty() def removeAllParameters(self): self._pycalendar.setParameters({}) self._markAsDirty() def removeParameterValue(self, paramname, paramvalue): paramname = paramname.upper() for attrName in self.parameterNames(): if attrName.upper() == paramname: for attr in tuple(self._pycalendar.getParameters()[attrName]): for value in attr.getValues(): if value == paramvalue: if not attr.removeValue(value): self._pycalendar.removeParameters(paramname) self._markAsDirty() def containsTimeRange(self, start, end, defaulttz=None): """ Determines whether this property contains a date/date-time within the specified start/end period. The only properties allowed for this query are: COMPLETED, CREATED, DTSTAMP and LAST-MODIFIED (caldav -09). @param start: a L{DateTime} specifying the beginning of the given time span. @param end: a L{DateTime} specifying the end of the given time span. C{end} may be None, indicating that there is no end date. @param defaulttz: the default L{PyTimezone} to use in datetime comparisons. @return: True if the property's date/date-time value is within the given time range, False if not, or the property is not an appropriate date/date-time value. """ # Verify that property name matches the ones allowed allowedNames = ["COMPLETED", "CREATED", "DTSTAMP", "LAST-MODIFIED"] if self.name() not in allowedNames: return False # get date/date-time value dt = self._pycalendar.getValue().getValue() assert isinstance(dt, DateTime), "Not a date/date-time value: {0!r}".format(self,) return timeRangesOverlap(dt, None, start, end, defaulttz) class Component (object): """ X{iCalendar} component. """ # Private Event access levels. ACCESS_PROPERTY = "X-CALENDARSERVER-ACCESS" ACCESS_PUBLIC = "PUBLIC" ACCESS_PRIVATE = "PRIVATE" ACCESS_CONFIDENTIAL = "CONFIDENTIAL" ACCESS_RESTRICTED = "RESTRICTED" accessMap = { "PUBLIC" : ACCESS_PUBLIC, "PRIVATE" : ACCESS_PRIVATE, "CONFIDENTIAL" : ACCESS_CONFIDENTIAL, "RESTRICTED" : ACCESS_RESTRICTED, } confidentialPropertiesMap = { "VCALENDAR": ("PRODID", "VERSION", "CALSCALE", ACCESS_PROPERTY), "VEVENT": ("UID", "RECURRENCE-ID", "SEQUENCE", "DTSTAMP", "STATUS", "TRANSP", "DTSTART", "DTEND", "DURATION", "RRULE", "RDATE", "EXRULE", "EXDATE",), "VTODO": ("UID", "RECURRENCE-ID", "SEQUENCE", "DTSTAMP", "STATUS", "DTSTART", "COMPLETED", "DUE", "DURATION", "RRULE", "RDATE", "EXRULE", "EXDATE",), "VJOURNAL": ("UID", "RECURRENCE-ID", "SEQUENCE", "DTSTAMP", "STATUS", "DTSTART", "RRULE", "RDATE", "EXRULE", "EXDATE",), "VFREEBUSY": ("UID", "DTSTAMP", "DTSTART", "DTEND", "DURATION", "FREEBUSY",), "VTIMEZONE": None, } extraRestrictedProperties = ("SUMMARY", "LOCATION",) # Hidden instance. HIDDEN_INSTANCE_PROPERTY = "X-CALENDARSERVER-HIDDEN-INSTANCE" allowedTypesList = None @classmethod def allowedTypes(cls): if cls.allowedTypesList is None: cls.allowedTypesList = ["text/calendar"] if config.EnableJSONData: cls.allowedTypesList.append("application/calendar+json") return cls.allowedTypesList @classmethod def allFromString(clazz, string, format=None): """ Just default to reading a single VCALENDAR """ return clazz.fromString(string, format) @classmethod def allFromStream(clazz, stream, format=None): """ Just default to reading a single VCALENDAR """ return clazz.fromStream(stream, format) @classmethod def fromString(clazz, string, format=None): """ Construct a L{Component} from a string. @param string: a string containing iCalendar data. @return: a L{Component} representing the first component described by C{string}. """ return clazz._fromData(string, False, format) @classmethod def fromStream(clazz, stream, format=None): """ Construct a L{Component} from a stream. @param stream: a C{read()}able stream containing iCalendar data. @return: a L{Component} representing the first component described by C{stream}. """ return clazz._fromData(stream, True, format) @classmethod def _fromData(clazz, data, isstream, format=None): """ Construct a L{Component} from a stream. @param stream: a C{read()}able stream containing iCalendar data. @param format: a C{str} indicating whether the data is iCalendar or jCal @return: a L{Component} representing the first component described by C{stream}. """ if isstream: pass else: if type(data) is unicode: data = data.encode("utf-8") else: # Valid utf-8 please data.decode("utf-8") # No BOMs please if data[:3] == codecs.BOM_UTF8: data = data[3:] errmsg = "Unknown" try: result = Calendar.parseData(data, format) except ErrorBase, e: errmsg = "{0}: {1}".format(e.mReason, e.mData,) result = None if not result: if isstream: data.seek(0) data = data.read() raise InvalidICalendarDataError("{0}\n{1}".format(errmsg, data,)) return clazz(None, pycalendar=result) @classmethod def fromIStream(clazz, stream, format=None): """ Construct a L{Component} from a stream. @param stream: an L{IStream} containing iCalendar data. @return: a deferred returning a L{Component} representing the first component described by C{stream}. """ # # FIXME: # This reads the request body into a string and then parses it. # A better solution would parse directly and incrementally from the # request stream. # def parse(data): return clazz.fromString(data, format) return allDataFromStream(IStream(stream), parse) @classmethod def newCalendar(cls): """ Create and return an empty C{VCALENDAR} component. @return: a new C{VCALENDAR} component with appropriate metadata properties already set (version, product ID). @rtype: an instance of this class """ self = cls("VCALENDAR") self.addProperty(Property("VERSION", "2.0")) self.addProperty(Property("PRODID", iCalendarProductID)) return self def __init__(self, name, **kwargs): """ Use this constructor to initialize an empty L{Component}. To create a new L{Component} from X{iCalendar} data, don't use this constructor directly. Use one of the factory methods instead. @param name: the name (L{str}) of the X{iCalendar} component type for the component. """ if name is None: if "pycalendar" in kwargs: pyobj = kwargs["pycalendar"] if pyobj is not None: if not isinstance(pyobj, ComponentBase): raise TypeError("Not a ComponentBase: {0!r}".format(pyobj,)) self._pycalendar = pyobj else: raise AssertionError("name may not be None") if "parent" in kwargs: parent = kwargs["parent"] if parent is not None: if not isinstance(parent, Component): raise TypeError("Not a Component: {0!r}".format(parent,)) self._parent = parent else: self._parent = None else: # FIXME: figure out creating an arbitrary component self._pycalendar = Calendar(add_defaults=False) if name == "VCALENDAR" else PyComponent.makeComponent(name, None) self._parent = None def __str__(self): """ NB This does not automatically include timezones in VCALENDAR objects. """ cachedCopy = getattr(self, "_cachedCopy", None) if cachedCopy is not None: return cachedCopy self._cachedCopy = str(self._pycalendar) return self._cachedCopy def _markAsDirty(self): """ Invalidate the cached copy of serialized icalendar data """ self._cachedCopy = None parent = getattr(self, "_parent", None) if parent is not None: parent._markAsDirty() def __repr__(self): return ( "<{self.__class__.__name__}: {pycal!r}>" .format(self=self, pycal=str(self._pycalendar)) ) def __hash__(self): return hash(str(self)) def __ne__(self, other): return not self.__eq__(other) def __eq__(self, other): if not isinstance(other, Component): return False return self._pycalendar == other._pycalendar def getText(self, format=None): return self.getTextWithTimezones(False, format) def getTextWithTimezones(self, includeTimezones, format=None): """ Return text representation and include timezones if the option is on. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) result = self._pycalendar.getText(includeTimezones=includeTimezones, format=format) if result is None: raise ValueError("Unknown format requested for calendar data.") return result # FIXME: Should this not be in __eq__? def same(self, other): return self._pycalendar == other._pycalendar def name(self): """ @return: the name of the iCalendar type of this component. """ return self._pycalendar.getType() def mainType(self): """ Determine the primary type of iCal component in this calendar. @return: the name of the primary type. @raise: L{InvalidICalendarDataError} if there is more than one primary type. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) mtype = None for component in self.subcomponents(): if component.name() in ignoredComponents: continue elif mtype and (mtype != component.name()): raise InvalidICalendarDataError("Component contains more than one type of primary type: {0!r}".format(self,)) else: mtype = component.name() return mtype def mainComponent(self): """ Return the primary iCal component in this calendar. If a master component exists, use that, otherwise use the first override. @return: the L{Component} of the primary type. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) result = None for component in self.subcomponents(): if component.name() in ignoredComponents: continue if not component.hasProperty("RECURRENCE-ID"): return component elif result is None: result = component return result def masterComponent(self): """ Return the master iCal component in this calendar. @return: the L{Component} for the master component, or C{None} if there isn't one. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) for component in self.subcomponents(): if component.name() in ignoredComponents: continue if not component.hasProperty("RECURRENCE-ID"): return component return None def overriddenComponent(self, recurrence_id): """ Return the overridden iCal component in this calendar matching the supplied RECURRENCE-ID property. This also returns the matching master component if recurrence_id is C{None}. @param recurrence_id: The RECURRENCE-ID property value to match. @type recurrence_id: L{DateTime} @return: the L{Component} for the overridden component, or C{None} if there isn't one. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) if isinstance(recurrence_id, str): recurrence_id = DateTime.parseText(recurrence_id) if recurrence_id else None for component in self.subcomponents(): if component.name() in ignoredComponents: continue rid = component.getRecurrenceIDUTC() if rid and recurrence_id and rid == recurrence_id: return component elif rid is None and recurrence_id is None: return component return None def accessLevel(self, default=ACCESS_PUBLIC): """ Return the access level for this component. @return: the access level for the calendar data. """ assert self.name() == "VCALENDAR", "Must be a VCALENDAR: {0!r}".format(self,) access = self.propertyValue(Component.ACCESS_PROPERTY) if access: access = access.upper() return Component.accessMap.get(access, default) def duplicate(self): """ Duplicate this object and all its contents. @return: the duplicated calendar. """ result = Component(None, pycalendar=self._pycalendar.duplicate()) if hasattr(self, "noInstanceIndexing"): result.noInstanceIndexing = self.noInstanceIndexing return result def subcomponents(self): """ @return: an iterable of L{Component} objects, one for each subcomponent of this component. """ return ( Component(None, pycalendar=c, parent=self) for c in self._pycalendar.getComponents() ) def addComponent(self, component): """ Adds a subcomponent to this component. @param component: the L{Component} to add as a subcomponent of this component. """ self._pycalendar.addComponent(component._pycalendar) component._parent = self self._markAsDirty() def removeComponent(self, component): """ Removes a subcomponent from this component. @param component: the L{Component} to remove. """ self._pycalendar.removeComponent(component._pycalendar) component._parent = None self._markAsDirty() def hasProperty(self, name): """ @param name: the name of the property whose existence is being tested. @return: True if the named property exists, False otherwise. """ return self._pycalendar.hasProperty(name) def getProperty(self, name): """ Get one property from the property list. @param name: the name of the property to get. @return: the L{Property} found or None. @raise: L{InvalidICalendarDataError} if there is more than one property of the given name. """ properties = tuple(self.properties(name)) if len(properties) == 1: return properties[0] if len(properties) > 1: raise InvalidICalendarDataError("More than one {0} property in component {1!r}".format(name, self)) return None def properties(self, name=None): """ @param name: if given and not C{None}, restricts the returned properties to those with the given C{name}. @return: an iterable of L{Property} objects, one for each property of this component. """ properties = [] if name is None: [properties.extend(i) for i in self._pycalendar.getProperties().values()] elif self._pycalendar.countProperty(name) > 0: properties = self._pycalendar.getProperties(name) return ( Property(None, None, None, parent=self, pycalendar=p) for p in properties ) def propertyValue(self, name): properties = tuple(self.properties(name)) if len(properties) == 1: return properties[0].value() if len(properties) > 1: raise InvalidICalendarDataError("More than one {0} property in component {1!r}".format(name, self)) return None def getStartDateUTC(self): """ Return the start date or date-time for the specified component converted to UTC. @param component: the Component whose start should be returned. @return: the L{DateTime} for the start. """ dtstart = self.propertyValue("DTSTART") return dtstart.duplicateAsUTC() if dtstart is not None else None def getEndDateUTC(self): """ Return the end date or date-time for the specified component, taking into account the presence or absence of DTEND/DURATION properties. The returned date-time is converted to UTC. @param component: the Component whose end should be returned. @return: the L{DateTime} for the end. """ dtend = self.propertyValue("DTEND") if dtend is None: dtstart = self.propertyValue("DTSTART") duration = self.propertyValue("DURATION") if duration is not None: dtend = dtstart + duration return dtend.duplicateAsUTC() if dtend is not None else None def getDueDateUTC(self): """ Return the due date or date-time for the specified component converted to UTC. Use DTSTART/DURATION if no DUE property. @param component: the Component whose start should be returned. @return: the L{DateTime} for the start. """ due = self.propertyValue("DUE") if due is None: dtstart = self.propertyValue("DTSTART") duration = self.propertyValue("DURATION") if dtstart is not None and duration is not None: due = dtstart + duration return due.duplicateAsUTC() if due is not None else None def getCompletedDateUTC(self): """ Return the completed date or date-time for the specified component converted to UTC. @param component: the Component whose start should be returned. @return: the datetime.date or datetime.datetime for the start. """ completed = self.propertyValue("COMPLETED") return completed.duplicateAsUTC() if completed is not None else None def getCreatedDateUTC(self): """ Return the created date or date-time for the specified component converted to UTC. @param component: the Component whose start should be returned. @return: the datetime.date or datetime.datetime for the start. """ created = self.propertyValue("CREATED") return created.duplicateAsUTC() if created is not None else None def getRecurrenceIDUTC(self): """ Return the recurrence-id for the specified component. @param component: the Component whose r-id should be returned. @return: the L{DateTime} for the r-id. """ rid = self.propertyValue("RECURRENCE-ID") return rid.duplicateAsUTC() if rid is not None else None def getRange(self): """ Determine whether a RANGE=THISANDFUTURE parameter is present on any RECURRENCE-ID property. @return: True if the parameter is present, False otherwise. """ ridprop = self.getProperty("RECURRENCE-ID") if ridprop is not None: range = ridprop.parameterValue("RANGE") if range is not None: return (range == "THISANDFUTURE") return False def getExdates(self): """ Get the set of all EXDATEs in this (master) component. """ exdates = set() for property in self.properties("EXDATE"): for exdate in property.value(): exdates.add(exdate.getValue()) return exdates def getTriggerDetails(self): """ Return the trigger information for the specified alarm component. @param component: the Component whose start should be returned. @return: a tuple consisting of: trigger : the 'native' trigger value related : either True (for START) or False (for END) repeat : an integer for the REPEAT count duration: the repeat duration if present, otherwise None """ assert self.name() == "VALARM", "Component is not a VAlARM: {0!r}".format(self,) # The trigger value trigger = self.propertyValue("TRIGGER") if trigger is None: raise InvalidICalendarDataError("VALARM has no TRIGGER property: {0!r}".format(self,)) # The related parameter related = self.getProperty("TRIGGER").parameterValue("RELATED") if related is None: related = True else: related = (related == "START") # Repeat property repeat = self.propertyValue("REPEAT") if repeat is None: repeat = 0 else: repeat = int(repeat) # Duration property duration = self.propertyValue("DURATION") if repeat > 0 and duration is None: raise InvalidICalendarDataError("VALARM has invalid REPEAT/DURATIOn properties: {0!r}".format(self,)) return (trigger, related, repeat, duration) def getRecurrenceSet(self): return self._pycalendar.getRecurrenceSet() def getEffectiveStartEnd(self): # Get the start/end range needed for instance comparisons if self.name() in ("VEVENT", "VJOURNAL",): return self.getStartDateUTC(), self.getEndDateUTC() elif self.name() == "VTODO": start = self.getStartDateUTC() due = self.getDueDateUTC() if start is None and due is not None: return due, due else: return start, due else: return None, None def getFBType(self): # Only VEVENTs block time if self.name() not in ("VEVENT",): return "FREE" # Handle status status = self.propertyValue("STATUS") if status == "CANCELLED": return "FREE" elif status == "TENTATIVE": return "BUSY-TENTATIVE" else: return "BUSY" def addProperty(self, property): """ Adds a property to this component. @param property: the L{Property} to add to this component. """ self._pycalendar.addProperty(property._pycalendar) self._pycalendar.finalise() property._parent = self self._markAsDirty() def removeProperty(self, property): """ Remove a property from this component. @param property: the L{Property} to remove from this component. """ if isinstance(property, str): for property in self.properties(property): self.removeProperty(property) else: self._pycalendar.removeProperty(property._pycalendar) self._pycalendar.finalise() property._parent = None self._markAsDirty() def removeAllPropertiesWithName(self, pname): """ Remove all properties with the given name from all components. @param pname: the property name to remove from all components. @type pname: C{str} """ for property in self.properties(pname): self.removeProperty(property) for component in self.subcomponents(): component.removeAllPropertiesWithName(pname) def replaceProperty(self, property): """ Add or replace a property in this component. @param property: the L{Property} to add or replace in this component. """ # Remove all existing ones first self._pycalendar.removeProperties(property.name()) self.addProperty(property) self._markAsDirty() def timezoneIDs(self): """ Returns the set of TZID parameter values appearing in any property in this component. @return: a set of strings, one for each unique TZID value. """ result = set() if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() != "VTIMEZONE": result.update(component.timezoneIDs()) else: for property in self.properties(): tzid = property.parameterValue("TZID") if tzid is not None: result.add(tzid) break return result def timezones(self): """ Returns the set of TZID's for each VTIMEZONE component. @return: a set of strings, one for each unique TZID value. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) results = set() for component in self.subcomponents(): if component.name() == "VTIMEZONE": results.add(component.propertyValue("TZID")) return results def truncateRecurrence(self, maximumCount): """ Truncate RRULEs etc to make sure there are no more than the given number of instances. @param maximumCount: the maximum number of instances to allow @type maximumCount: C{int} @return: a C{bool} indicating whether a change was made or not """ changed = False master = self.masterComponent() if master and master.isRecurring(): rrules = master._pycalendar.getRecurrenceSet() if rrules: for rrule in rrules.getRules(): if rrule.getUseCount(): # Make sure COUNT is less than the limit if rrule.getCount() > maximumCount: rrule.setCount(maximumCount) changed = True elif rrule.getUseUntil(): # Need to figure out how to determine number of instances # with this UNTIL and truncate if needed start = master.getStartDateUTC() diff = differenceDateTime(start, rrule.getUntil()) diff = diff.getDays() * 24 * 60 * 60 + diff.getSeconds() period = { definitions.eRecurrence_YEARLY: 365 * 24 * 60 * 60, definitions.eRecurrence_MONTHLY: 30 * 24 * 60 * 60, definitions.eRecurrence_WEEKLY: 7 * 24 * 60 * 60, definitions.eRecurrence_DAILY: 1 * 24 * 60 * 60, definitions.eRecurrence_HOURLY: 60 * 60, definitions.eRecurrence_MINUTELY: 60, definitions.eRecurrence_SECONDLY: 1 }[rrule.getFreq()] * rrule.getInterval() if diff / period > maximumCount: rrule.setUseUntil(False) rrule.setUseCount(True) rrule.setCount(maximumCount) rrules.changed() changed = True else: # For frequencies other than yearly we will truncate at our limit if rrule.getFreq() != definitions.eRecurrence_YEARLY: rrule.setUseCount(True) rrule.setCount(maximumCount) rrules.changed() changed = True if changed: self._markAsDirty() return changed def onlyPastInstances(self, rid): """ Remove all recurrence instances at or beyond the specified recurrence-id. Adjust the bounds of any RRULEs to match the new limit, remove RDATEs/EXDATEs and overridden components beyond the limit. @param rid: the recurrence-id limit @type rid: L{DateTime} """ if not self.isRecurring(): return master = self.masterComponent() if master: # Adjust any RRULE first rrules = master._pycalendar.getRecurrenceSet() if rrules: for rrule in rrules.getRules(): rrule.setUseUntil(True) rrule.setUseCount(False) until = rid.duplicate() until.offsetSeconds(-1) rrule.setUntil(until) # Remove any RDATEs or EXDATEs in the future for property in list(itertools.chain( master.properties("RDATE"), master.properties("EXDATE"), )): for value in list(property.value()): if value.getValue() >= rid: property.value().remove(value) if len(property.value()) == 0: master.removeProperty(property) # Remove overrides in the future for component in list(self.subcomponents()): if component.name() in ignoredComponents: continue c_rid = component.getRecurrenceIDUTC() if c_rid is not None and c_rid >= rid: self.removeComponent(component) # Handle per-user data component by removing ones in the future for component in list(self.subcomponents()): if component.name() == PERUSER_COMPONENT: for subcomponent in list(component.subcomponents()): c_rid = subcomponent.getRecurrenceIDUTC() if c_rid is not None and c_rid >= rid: component.removeComponent(subcomponent) if len(list(component.subcomponents())) == 0: self.removeComponent(component) self._markAsDirty() # We changed the instance set so remove any instance cache # TODO: we could be smarter here and truncate the instance list if hasattr(self, "cachedInstances"): delattr(self, "cachedInstances") def onlyFutureInstances(self, rid): """ Remove all recurrence instances from the specified recurrence-id into the past. Adjust the bounds of any RRULEs to match the new limit, remove RDATEs/EXDATEs and overridden components beyond the limit. This also requires "re-basing" the master component to the new first instance - but noting that has to match any RRULE pattern. @param rid: the recurrence-id limit @type rid: L{DateTime} """ if not self.isRecurring(): return master = self.masterComponent() if master: # Check if cut-off matches an RDATE adjusted_rid = rid adjust_rrule = None adjust_count = 0 continuing_rrule = True # Need to detect the first valid RRULE instance after the cut-off as that needs to be the new DTSTART rrules = master._pycalendar.getRecurrenceSet() if rrules and len(rrules.getRules()) != 0: rrule = rrules.getRules()[0] upperlimit = rid.duplicate() upperlimit.offsetYear(1) rrule_expanded = [] rrule.expand( master.propertyValue("DTSTART"), Period(DateTime(1900, 1, 1), upperlimit), rrule_expanded, ) for ctr, i in enumerate(sorted(rrule_expanded)): if i >= rid: adjusted_rid = i adjust_rrule = rrule adjust_count = ctr break else: # RRULE not needed in derived master continuing_rrule = False # Adjust master to previously derived instance derived = self.deriveInstance(adjusted_rid, allowExcluded=True) if derived is None: return # Adjust any COUNT to exclude the earlier instances - note we do this after # deriving the instance otherwise it might truncate the instance we care about if adjust_rrule is not None and rrule.getUseCount(): adjust_rrule.setCount(adjust_rrule.getCount() - adjust_count) # Fix up recurrence properties so the derived one looks like the master derived.removeProperty(derived.getProperty("RECURRENCE-ID")) for property in list(itertools.chain( master.properties("RRULE") if continuing_rrule else (), master.properties("RDATE"), master.properties("EXDATE"), )): derived.addProperty(property) # Now switch over to using the new "derived" master self.removeComponent(master) master = derived self.addComponent(master) # Remove any RDATEs or EXDATEs in the past for property in list(itertools.chain( master.properties("RDATE"), master.properties("EXDATE"), )): for value in list(property.value()): # If the derived master was derived from an RDATE we remove the RDATE if value.getValue() < rid or property.name() == "RDATE" and value.getValue() == adjusted_rid: property.value().remove(value) if len(property.value()) == 0: master.removeProperty(property) # Remove overrides in the past - but do not remove any override matching # the cut-off as that is still a valid override after "re-basing" the master. for component in list(self.subcomponents()): if component.name() in ignoredComponents: continue c_rid = component.getRecurrenceIDUTC() if c_rid is not None and c_rid < rid: self.removeComponent(component) # Handle per-user data component by removing ones in the past for component in list(self.subcomponents()): if component.name() == PERUSER_COMPONENT: for subcomponent in list(component.subcomponents()): c_rid = subcomponent.getRecurrenceIDUTC() if c_rid is not None and c_rid < rid: component.removeComponent(subcomponent) if len(list(component.subcomponents())) == 0: self.removeComponent(component) self._markAsDirty() # We changed the instance set so remove any instance cache # TODO: we could be smarter here and truncate the instance list if hasattr(self, "cachedInstances"): delattr(self, "cachedInstances") def expand(self, start, end, timezone=None): """ Expand the components into a set of new components, one for each instance in the specified range. Date-times are converted to UTC. A new calendar object is returned. @param start: the L{DateTime} for the start of the range. @param end: the L{DateTime} for the end of the range. @param timezone: the L{Component} or L{Timezone} of the VTIMEZONE to use for floating/all-day. @return: the L{Component} for the new calendar with expanded instances. """ if timezone is not None and isinstance(timezone, Component): pytz = Timezone(tzid=timezone.propertyValue("TZID")) else: pytz = timezone # Create new calendar object with same properties as the original, but # none of the originals sub-components calendar = Component("VCALENDAR") for property in calendar.properties(): calendar.removeProperty(property) for property in self.properties(): calendar.addProperty(property) # Expand the instances and add each one - use the normalizeForExpand date/time normalization method here # so that all-day date/times remain that way. However, when doing the timeRangesOverlap test below, we # Need to convert the all-days to floating (T000000) so that the timezone overlap calculation can be done # properly. instances = self.expandTimeRanges(end, lowerLimit=start, normalizeFunction=normalizeForExpand) first = True for key in instances: instance = instances[key] if timeRangesOverlap(normalizeForIndex(instance.start), normalizeForIndex(instance.end), start, end, pytz): calendar.addComponent(self.expandComponent(instance, first)) first = False return calendar def expandComponent(self, instance, first): """ Create an expanded component based on the instance provided. NB Expansion also requires UTC conversions. @param instance: an L{Instance} for the instance being expanded. @return: a new L{Component} for the expanded instance. """ # Duplicate the component from the instance newcomp = instance.component.duplicate() # Strip out unwanted recurrence properties for property in tuple(newcomp.properties()): if property.name() in ["RRULE", "RDATE", "EXRULE", "EXDATE", "RECURRENCE-ID"]: newcomp.removeProperty(property) # Convert all datetime properties to UTC unless they are floating for property in newcomp.properties(): value = property.value() if isinstance(value, DateTime) and value.local(): property.removeParameter("TZID") property.setValue(value.duplicateAsUTC()) # Now reset DTSTART, DTEND/DURATION for property in newcomp.properties("DTSTART"): property.setValue(instance.start) for property in newcomp.properties("DTEND"): property.setValue(instance.end) for property in newcomp.properties("DURATION"): property.setValue(instance.end - instance.start) # Add RECURRENCE-ID if not master instance if not instance.isMasterInstance(): newcomp.addProperty(Property("RECURRENCE-ID", instance.rid)) return newcomp def cacheExpandedTimeRanges(self, limit, ignoreInvalidInstances=False): """ Expand instances up to the specified limit and cache the results in this object so we can return cached results in the future. The limit value is the actual value that the requester needs, but we will cache an addition 365-days worth to give us some breathing room to return results for future instances. @param limit: the max datetime to cache up to. @type limit: L{DateTime} """ # Checked for cached values first if hasattr(self, "cachedInstances"): cachedLimit = self.cachedInstances.limit if cachedLimit is None or cachedLimit > limit: # We have already fully expanded, or cached up to the requested time, # so return cached instances return self.cachedInstances lookAheadLimit = limit + Duration(days=365) self.cachedInstances = self.expandTimeRanges( lookAheadLimit, ignoreInvalidInstances=ignoreInvalidInstances ) return self.cachedInstances def expandTimeRanges(self, limit, lowerLimit=None, ignoreInvalidInstances=False, normalizeFunction=normalizeForIndex): """ Expand the set of recurrence instances for the components contained within this VCALENDAR component. We will assume that this component has already been validated as a CalDAV resource (i.e. only one type of component, all with the same UID) @param limit: L{DateTime} value representing the end of the expansion. @param ignoreInvalidInstances: C{bool} whether to ignore instance errors. @return: a set of Instances for each recurrence in the set. """ componentSet = self.subcomponents() return self.expandSetTimeRanges(componentSet, limit, lowerLimit=lowerLimit, ignoreInvalidInstances=ignoreInvalidInstances, normalizeFunction=normalizeFunction) def expandSetTimeRanges(self, componentSet, limit, lowerLimit=None, ignoreInvalidInstances=False, normalizeFunction=normalizeForIndex): """ Expand the set of recurrence instances up to the specified date limit. What we do is first expand the master instance into the set of generate instances. Then we merge the overridden instances, taking into account THISANDFUTURE and THISANDPRIOR. @param componentSet: the set of components that are to make up the recurrence set. These MUST all be components with the same UID and type, forming a proper recurring set. @param limit: L{DateTime} value representing the end of the expansion. @param componentSet: the set of components that are to make up the recurrence set. These MUST all be components with the same UID and type, forming a proper recurring set. @type componentSet: C{list} @param limit: the end of the expansion @type limit: L{DateTime} @param ignoreInvalidInstances: whether or not invalid recurrences raise an exception @type ignoreInvalidInstances: C{bool} @param normalizeFunction: a function used to normalize date/time values in instances @type normalizeFunction: C{function} @return: L{InstanceList} containing expanded L{Instance} for each recurrence in the set. """ # Set of instances to return instances = InstanceList(ignoreInvalidInstances=ignoreInvalidInstances, normalizeFunction=normalizeFunction) try: instances.expandTimeRanges(componentSet, limit, lowerLimit=lowerLimit) except InvalidOverriddenInstanceError as e: if accountingEnabledForCategory("Invalid Instance"): emitAccounting( "Invalid Instance", self.resourceUID().encode("base64")[:-1], "{}\n\n{}".format(str(e), str(self)), ) raise return instances def getComponentInstances(self): """ Get the R-ID value for each component. @return: a tuple of recurrence-ids """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": result = () for component in self.subcomponents(): if component.name() not in ignoredComponents: result += component.getComponentInstances() return result else: rid = self.getRecurrenceIDUTC() return (rid,) def isRecurring(self): """ Check whether any recurrence properties are present in any component. """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents and component.isRecurring(): return True else: for propname in ("RRULE", "RDATE", "EXDATE", "RECURRENCE-ID",): if self.hasProperty(propname): return True return False def isRecurringUnbounded(self): """ Check for unbounded recurrence. """ master = self.masterComponent() if master: rrules = master.properties("RRULE") for rrule in rrules: if not rrule.value().getUseCount() and not rrule.value().getUseUntil(): return True return False def deriveInstance(self, rid, allowCancelled=False, newcomp=None, allowExcluded=False): """ Derive an instance from the master component that has the provided RECURRENCE-ID, but with all other properties, components etc from the master. If the requested override is currently marked as an EXDATE in the existing master, allow an option whereby the override is added as STATUS:CANCELLED and the EXDATE removed. IMPORTANT: all callers of this method MUST check the return value for None. Never assume that a valid instance will be derived - no matter how much you think you understand iCalendar recurrence. There is always some new thing that will surprise you. @param rid: recurrence-id value @type rid: L{DateTime} or C{str} @param allowCancelled: whether to allow a STATUS:CANCELLED override @type allowCancelled: C{bool} @param allowExcluded: whether to derive an instance for an existing EXDATE @type allowExcluded: C{bool} @return: L{Component} for newly derived instance, or C{None} if not a valid override """ if allowCancelled and newcomp is not None: raise ValueError("Cannot re-use master component with allowCancelled") # Must have a master component master = self.masterComponent() if master is None: return None if isinstance(rid, str): rid = DateTime.parseText(rid) if rid else None # TODO: Check that the recurrence-id is a valid instance # For now we just check that there is no matching EXDATE didCancel = False matchedExdate = False for exdate in tuple(master.properties("EXDATE")): for exdateValue in exdate.value(): if exdateValue.getValue() == rid: if allowCancelled: exdate.value().remove(exdateValue) if len(exdate.value()) == 0: master.removeProperty(exdate) didCancel = True # We changed the instance set so remove any instance cache if hasattr(self, "cachedInstances"): delattr(self, "cachedInstances") break elif allowExcluded: matchedExdate = True break else: # Cannot derive from an existing EXDATE return None if not matchedExdate: # Check whether recurrence-id matches an RDATE - if so it is OK rdates = set() for rdate in master.properties("RDATE"): rdates.update([item.getValue().duplicateAsUTC() for item in rdate.value()]) if rid not in rdates: # Check whether we have a truncated RRULE rrules = master.properties("RRULE") if len(tuple(rrules)): instances = self.cacheExpandedTimeRanges(rid) instance_rid = normalizeForIndex(rid) if str(instance_rid) not in instances.instances: # No match to a valid RRULE instance return None else: # No RRULE and no match to an RDATE => error return None # If we were fed an already derived component, use that, otherwise make a new one if newcomp is None: newcomp = self.masterDerived() # New DTSTART is the RECURRENCE-ID we are deriving but adjusted to the # original DTSTART's localtime dtstart = newcomp.getProperty("DTSTART") if newcomp.hasProperty("DTEND"): dtend = newcomp.getProperty("DTEND") oldduration = dtend.value() - dtstart.value() newdtstartValue = rid.duplicate() if not dtstart.value().isDateOnly(): if dtstart.value().local(): newdtstartValue.adjustTimezone(dtstart.value().getTimezone()) else: newdtstartValue.setDateOnly(True) dtstart.setValue(newdtstartValue) if newcomp.hasProperty("DTEND"): dtend.setValue(newdtstartValue + oldduration) newcomp.replaceProperty(Property("RECURRENCE-ID", dtstart.value(), params={})) if didCancel: newcomp.replaceProperty(Property("STATUS", "CANCELLED")) # After creating/changing a component we need to do this to keep PyCalendar happy newcomp._pycalendar.finalise() return newcomp def masterDerived(self): """ Generate a component from the master instance that can be fed repeatedly to deriveInstance in the case where the result of deriveInstance is not going to be inserted into the component. This provides an optimization for avoiding unnecessary .duplicate() calls on the master for each deriveInstance. """ # Must have a master component master = self.masterComponent() if master is None: return None # Create the derived instance newcomp = master.duplicate() # Strip out unwanted recurrence properties for property in tuple(newcomp.properties()): if property.name() in ("RRULE", "RDATE", "EXRULE", "EXDATE", "RECURRENCE-ID",): newcomp.removeProperty(property) return newcomp def validInstances(self, rids, ignoreInvalidInstances=False): """ Test whether the specified recurrence-ids are valid instances in this event. @param rid: recurrence-id values @type rid: iterable @return: C{set} of valid rids """ valid = set() non_master_rids = [rid for rid in rids if rid is not None] if non_master_rids: # Pre-cache instance expansion up to the highest rid highest_rid = max(non_master_rids) self.cacheExpandedTimeRanges( highest_rid + Duration(days=1), ignoreInvalidInstances=ignoreInvalidInstances ) for rid in rids: if self.validInstance(rid, clear_cache=False, ignoreInvalidInstances=ignoreInvalidInstances): valid.add(rid) return valid def validInstance(self, rid, clear_cache=True, ignoreInvalidInstances=False): """ Test whether the specified recurrence-id is a valid instance in this event. @param rid: recurrence-id value @type rid: L{DateTime} @return: C{bool} """ if self.masterComponent() is None: return rid in set(self.getComponentInstances()) if rid is None: return True # Get expansion instances = self.cacheExpandedTimeRanges( rid, ignoreInvalidInstances=ignoreInvalidInstances ) new_rids = set([instances[key].rid for key in instances]) return rid in new_rids def resourceUID(self): """ @return: the UID of the subcomponents in this component. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) if not hasattr(self, "_resource_uid"): for subcomponent in self.subcomponents(): if subcomponent.name() not in ignoredComponents: self._resource_uid = subcomponent.propertyValue("UID") break else: self._resource_uid = None return self._resource_uid def newUID(self, newUID=None): """ Generate a new UID for all components in this VCALENDAR """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) newUID = str(uuid.uuid4()) if newUID is None else newUID self._pycalendar.changeUID(self.resourceUID(), newUID) self._resource_uid = newUID self._markAsDirty() return self._resource_uid def resourceType(self): """ @return: the name of the iCalendar type of the subcomponents in this component. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) if not hasattr(self, "_resource_type"): has_timezone = False for subcomponent in self.subcomponents(): name = subcomponent.name() if name == "VTIMEZONE": has_timezone = True elif subcomponent.name() in ignoredComponents: continue else: self._resource_type = name break else: if has_timezone: self._resource_type = "VTIMEZONE" else: raise InvalidICalendarDataError("No component type found for calendar component: {0!r}".format(self,)) return self._resource_type def stripKnownTimezones(self): """ Remove timezones that this server knows about """ changed = False for subcomponent in tuple(self.subcomponents()): if subcomponent.name() == "VTIMEZONE": tzid = subcomponent.propertyValue("TZID") try: hasTZ(tzid) except TimezoneException: # tzid not available - do not strip pass else: # tzid known - strip component out self.removeComponent(subcomponent) changed = True return changed def validCalendarData(self, doFix=True, doRaise=True, validateRecurrences=False): """ @return: tuple of fixed, unfixed issues @raise InvalidICalendarDataError: if the given calendar data is not valid and cannot be fixed. """ if self.name() != "VCALENDAR": log.debug("Not a calendar: {0}".format(self,)) raise InvalidICalendarDataError("Not a calendar") if not self.resourceType(): log.debug("Unknown resource type: {0}".format(self,)) raise InvalidICalendarDataError("Unknown resource type") # Do underlying iCalendar library validation with data fix fixed, unfixed = self._pycalendar.validate(doFix=doFix) # Detect invalid occurrences and fix by adding RDATEs for them if validateRecurrences: rfixed, runfixed = self.validRecurrenceIDs(doFix=doFix) fixed.extend(rfixed) unfixed.extend(runfixed) if unfixed: log.debug("Calendar data had unfixable problems:\n {0}".format("\n ".join(unfixed),)) if doRaise: raise InvalidICalendarDataError("Calendar data had unfixable problems:\n {0}".format("\n ".join(unfixed),)) if fixed: log.debug("Calendar data had fixable problems:\n {0}".format("\n ".join(fixed),)) return fixed, unfixed def validRecurrenceIDs(self, doFix=True): fixed = [] unfixed = [] # Detect invalid occurrences and fix by adding RDATEs for them master = self.masterComponent() if master is not None: # Get the set of all recurrence IDs all_rids = set(self.getComponentInstances()) if None in all_rids: all_rids.remove(None) # If the master has no recurrence properties treat any other components as invalid if master.isRecurring(): # Remove all EXDATEs with a matching RECURRENCE-ID. Do this before we start # processing of valid instances just in case the matching R-ID is also not valid and # thus will need RDATE added. exdates = {} for property in list(master.properties("EXDATE")): for exdate in property.value(): exdates[exdate.getValue()] = property for rid in all_rids: if rid in exdates: if doFix: property = exdates[rid] for value in property.value(): if value.getValue() == rid: property.value().remove(value) break master.removeProperty(property) if len(property.value()) > 0: master.addProperty(property) del exdates[rid] fixed.append("Removed EXDATE for valid override: {0}".format(rid,)) else: unfixed.append("EXDATE for valid override: {0}".format(rid,)) # Get the set of all valid recurrence IDs valid_rids = self.validInstances(all_rids, ignoreInvalidInstances=True) # Get the set of all RDATEs and add those to the valid set rdates = [] for property in master.properties("RDATE"): rdates.extend([_rdate.getValue() for _rdate in property.value()]) valid_rids.update(set(rdates)) # Remove EXDATEs predating master dtstart = master.propertyValue("DTSTART") if dtstart is not None: for property in list(master.properties("EXDATE")): newValues = [] changed = False for exdate in property.value(): exdateValue = exdate.getValue() if exdateValue < dtstart: if doFix: fixed.append("Removed earlier EXDATE: {0}".format(exdateValue,)) else: unfixed.append("EXDATE earlier than master: {0}".format(exdateValue,)) changed = True else: newValues.append(exdateValue) if changed and doFix: # Remove the property... master.removeProperty(property) if newValues: # ...and add it back only if it still has values property.setValue(newValues) master.addProperty(property) else: valid_rids = set() # Determine the invalid recurrence IDs by set subtraction invalid_rids = all_rids - valid_rids # Add RDATEs for the invalid ones, or remove any EXDATE. for invalid_rid in invalid_rids: brokenComponent = self.overriddenComponent(invalid_rid) brokenRID = brokenComponent.propertyValue("RECURRENCE-ID") if doFix: master.addProperty(Property("RDATE", [brokenRID, ])) fixed.append( "Added RDATE for invalid occurrence: {0}".format( brokenRID, ) ) else: unfixed.append("Invalid occurrence: {0}".format(brokenRID,)) return fixed, unfixed def validCalendarForCalDAV(self, methodAllowed): """ @param methodAllowed: True if METHOD property is allowed, False otherwise. @raise InvalidICalendarDataError: if the given calendar component is not valid for use as a X{CalDAV} resource. """ # Disallowed in CalDAV-Access-08, section 4.1 if not methodAllowed and self.hasProperty("METHOD"): msg = "METHOD property is not allowed in CalDAV iCalendar data" log.debug(msg) raise InvalidICalendarDataError(msg) # # Must not contain more than one type of iCalendar component, except for # the required timezone components, and component UIDs must match # ctype = None component_id = None component_rids = set() timezone_refs = set() timezones = set() got_master = False # got_override = False # master_recurring = False for subcomponent in self.subcomponents(): if subcomponent.name() == "VTIMEZONE": timezones.add(subcomponent.propertyValue("TZID")) elif subcomponent.name() in ignoredComponents: continue else: if ctype is None: ctype = subcomponent.name() else: if ctype != subcomponent.name(): msg = "Calendar resources may not contain more than one type of calendar component ({0} and {1} found)".format( ctype, subcomponent.name() ) log.debug(msg) raise InvalidICalendarDataError(msg) if ctype not in allowedComponents: msg = "Component type: {0} not allowed".format(ctype,) log.debug(msg) raise InvalidICalendarDataError(msg) uid = subcomponent.propertyValue("UID") if uid is None: msg = "All components must have UIDs" log.debug(msg) raise InvalidICalendarDataError(msg) rid = subcomponent.getRecurrenceIDUTC() # Verify that UIDs are the same if component_id is None: component_id = uid elif component_id != uid: msg = "Calendar resources may not contain components with different UIDs ({0} and {1} found)".format( component_id, subcomponent.propertyValue("UID") ) log.debug(msg) raise InvalidICalendarDataError(msg) # Verify that there is only one master component if rid is None: if got_master: msg = "Calendar resources may not contain components with the same UIDs and no Recurrence-IDs ({0} and {1} found)".format( component_id, subcomponent.propertyValue("UID") ) log.debug(msg) raise InvalidICalendarDataError(msg) else: got_master = True # master_recurring = subcomponent.hasProperty("RRULE") or subcomponent.hasProperty("RDATE") else: pass # got_override = True # Check that if an override is present then the master is recurring # Leopard iCal sometimes does this for overridden instances that an Attendee receives and # it creates a "fake" (invalid) master. We are going to skip this test here. Instead implicit # scheduling will verify the validity of the components and raise if they don't make sense. # If no scheduling is happening then we allow this - that may cause other clients to choke. # If it does we will have to reinstate this check but only after we have checked for implicit. # UNCOMMENT OUT master_recurring AND got_override ASSIGNMENTS ABOVE # if got_override and got_master and not master_recurring: # msg = "Calendar resources must have a recurring master component if there is an overridden one (%s)" % (subcomponent.propertyValue("UID"),) # log.debug(msg) # raise InvalidICalendarDataError(msg) # Check for duplicate RECURRENCE-IDs if rid in component_rids: msg = "Calendar resources may not contain components with the same Recurrence-IDs ({0})".format(rid,) log.debug(msg) raise InvalidICalendarDataError(msg) else: component_rids.add(rid) timezone_refs.update(subcomponent.timezoneIDs()) # # Make sure required timezone components are present # if not config.EnableTimezonesByReference: for timezone_ref in timezone_refs: if timezone_ref not in timezones: msg = "Timezone ID {0} is referenced but not defined: {1}".format(timezone_ref, self,) log.debug(msg) raise InvalidICalendarDataError(msg) # # FIXME: # This test is not part of the spec; it appears to be legal (but # goofy?) to have extra timezone components. # for timezone in timezones: if timezone not in timezone_refs: log.debug( "Timezone {0} is not referenced by any non-timezone component".format(timezone,) ) # TZIDs without a VTIMEZONE must be available in the server's TZ database missing_timezones = timezone_refs - timezones for tzid in missing_timezones: # Will raise TimezoneException if tzid not present in server's database hasTZ(tzid) # Control character check - only HTAB, CR, LF allowed for characters in the range 0x00-0x1F s = str(self) if len(s.translate(None, "\x00\x01\x02\x03\x04\x05\x06\x07\x08\x0B\x0C\x0E\x0F\x10\x11\x12\x13\x14\x15\x16\x17\x18\x19\x1A\x1B\x1C\x1D\x1E\x1F")) != len(s): raise InvalidICalendarDataError("iCalendar contains illegal control character") def validOrganizerForScheduling(self, doFix=True): """ Check that the ORGANIZER property is valid for scheduling """ organizers = self.getOrganizersByInstance() foundOrganizer = None foundRid = None missingRids = set() for organizer, rid in organizers: if organizer: if foundOrganizer: if organizer != foundOrganizer: # We have different ORGANIZERs in the same iCalendar object - this is an error msg = "Only one ORGANIZER is allowed in an iCalendar object:\n{0}".format(self,) log.debug(msg) raise InvalidICalendarDataError(msg) else: foundOrganizer = organizer foundRid = rid else: missingRids.add(rid) # If there are some components without an ORGANIZER we will fix the data if foundOrganizer and missingRids: if doFix: log.debug("Fixing missing ORGANIZER properties") organizerProperty = self.overriddenComponent(foundRid).getProperty("ORGANIZER") for rid in missingRids: self.overriddenComponent(rid).addProperty(organizerProperty) else: raise InvalidICalendarDataError("iCalendar missing ORGANIZER properties in some instances") return foundOrganizer def gettimezone(self): """ Get the Timezone for a Timezone component. @return: L{Timezone} if this is a VTIMEZONE, otherwise None. """ if self.name() == "VTIMEZONE": return Timezone(tzid=self._pycalendar.getID()) elif self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() == "VTIMEZONE": return component.gettimezone() else: return None else: return None # # # iTIP stuff # # def isValidMethod(self): """ Verify that this calendar component has a valid iTIP METHOD property. @return: True if valid, False if not """ try: method = self.propertyValue("METHOD") if method not in ("PUBLISH", "REQUEST", "REPLY", "ADD", "CANCEL", "REFRESH", "COUNTER", "DECLINECOUNTER"): return False except InvalidICalendarDataError: return False return True def isValidITIP(self): """ Verify that this calendar component is valid according to iTIP. @return: True if valid, False if not """ try: method = self.propertyValue("METHOD") if method not in ("PUBLISH", "REQUEST", "REPLY", "ADD", "CANCEL", "REFRESH", "COUNTER", "DECLINECOUNTER"): return False # First make sure components are all of the same time (excluding VTIMEZONE) self.validCalendarForCalDAV(methodAllowed=True) # Next we could check the iTIP status for each type of method/component pair, however # we can also leave that up to the server except for the REQUEST/VFREEBUSY case which # the server will handle itself. if (method == "REQUEST") and (self.mainType() == "VFREEBUSY"): # TODO: verify REQUEST/VFREEBUSY as being OK # Only one VFREEBUSY (actually multiple X-'s are allowed but we will reject) if len([c for c in self.subcomponents()]) != 1: return False except InvalidICalendarDataError: return False return True def getOrganizer(self): """ Get the organizer value. Works on either a VCALENDAR or on a component. @return: the string value of the Organizer property, or None """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: return component.getOrganizer() else: try: # Find the primary subcomponent return self.propertyValue("ORGANIZER") except InvalidICalendarDataError: pass return None def getOrganizersByInstance(self): """ Get the organizer value for each instance. @return: a list of tuples of (organizer value, recurrence-id) """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": result = () for component in self.subcomponents(): if component.name() not in ignoredComponents: result += component.getOrganizersByInstance() return result else: try: # Should be just one ORGANIZER org = self.propertyValue("ORGANIZER") rid = self.getRecurrenceIDUTC() return ((org, rid),) except InvalidICalendarDataError: pass return () def getOrganizerProperty(self): """ Get the organizer value. Works on either a VCALENDAR or on a component. @return: the string value of the Organizer property, or None """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: return component.getOrganizerProperty() else: try: # Find the primary subcomponent return self.getProperty("ORGANIZER") except InvalidICalendarDataError: pass return None def getOrganizerScheduleAgent(self): is_server = False organizerProp = self.getOrganizerProperty() if organizerProp is not None: if organizerProp.hasParameter("SCHEDULE-AGENT"): if organizerProp.parameterValue("SCHEDULE-AGENT") == "SERVER": is_server = True else: is_server = True return is_server def cleanOrganizerScheduleAgent(self): """ Remove components whose ORGANIZER property does not have SCHEDULE-AGENT=SERVER. """ changed = False for component in tuple(self.subcomponents()): if component.name() in ignoredComponents: continue organizerProp = component.getOrganizerProperty() if organizerProp is not None: if organizerProp.parameterValue("SCHEDULE-AGENT", "SERVER") != "SERVER": self.removeComponent(component) changed = True return changed def recipientPropertyName(self): return "VOTER" if self.name() == "VPOLL" else "ATTENDEE" def getAttendees(self): """ Get the attendee value. Works on either a VCALENDAR or on a component. @param match: a C{list} of calendar user address strings to try and match. @return: a C{list} of the string values of the Attendee property, or None """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: return component.getAttendees() else: # Find the property values return [p.value() for p in self.properties(self.recipientPropertyName())] return None def getAttendeesByInstance(self, makeUnique=False, onlyScheduleAgentServer=False): """ Get the attendee values for each instance. Optionally remove duplicates. @param makeUnique: if C{True} remove duplicate ATTENDEEs in each component @type makeUnique: C{bool} @param onlyScheduleAgentServer: if C{True} only return ATETNDEEs with SCHEDULE-AGENT=SERVER set @type onlyScheduleAgentServer: C{bool} @return: a list of tuples of (organizer value, recurrence-id) """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": result = () for component in self.subcomponents(): if component.name() not in ignoredComponents: result += component.getAttendeesByInstance(makeUnique, onlyScheduleAgentServer) return result else: result = () attendees = set() rid = self.getRecurrenceIDUTC() for attendee in tuple(self.properties(self.recipientPropertyName())): if onlyScheduleAgentServer: if attendee.hasParameter("SCHEDULE-AGENT"): if attendee.parameterValue("SCHEDULE-AGENT") != "SERVER": continue cuaddr = attendee.value() if makeUnique and cuaddr in attendees: self.removeProperty(attendee) else: result += ((cuaddr, rid),) attendees.add(cuaddr) return result def getVoterProperty(self, match): """ Get the voters matching a value. @param match: a C{list} of calendar user address strings to try and match. @return: the matching Voter property, or None """ # Need to normalize http/https cu addresses test = set() for item in match: test.add(normalizeCUAddr(item)) # Find the primary subcomponent for voter in self.properties("VOTER"): if normalizeCUAddr(voter.value()) in test: return voter return None def getAttendeeProperty(self, match): """ Get the attendees matching a value. Works on either a VCALENDAR or on a component. @param match: a C{list} of calendar user address strings to try and match. @return: the matching Attendee property, or None """ # Need to normalize http/https cu addresses test = set() for item in match: test.add(normalizeCUAddr(item)) # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: attendee = component.getAttendeeProperty(match) if attendee is not None: return attendee else: # Find the primary subcomponent for attendee in self.properties(self.recipientPropertyName()): if normalizeCUAddr(attendee.value()) in test: return attendee return None def getAttendeeProperties(self, match): """ Get all the attendees matching a value in each component. Works on a VCALENDAR component only. @param match: a C{list} of calendar user address strings to try and match. @return: the string value of the Organizer property, or None """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) # Extract appropriate sub-component if this is a VCALENDAR results = [] for component in self.subcomponents(): if component.name() not in ignoredComponents: attendee = component.getAttendeeProperty(match) if attendee: results.append(attendee) return results def getAllAttendeeProperties(self): """ Yield all attendees as Property objects. Works on either a VCALENDAR or on a component. @return: a generator yielding Property objects """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: for attendee in component.getAllAttendeeProperties(): yield attendee else: # Find the primary subcomponent for attendee in self.properties(self.recipientPropertyName()): yield attendee def getAllUniqueAttendees(self, onlyScheduleAgentServer=True): attendeesByInstance = self.getAttendeesByInstance(True, onlyScheduleAgentServer=onlyScheduleAgentServer) attendees = set() for attendee, _ignore in attendeesByInstance: attendees.add(attendee) return attendees def getMaskUID(self): """ Get the X-CALENDARSEREVR-MASK-UID value. Works on either a VCALENDAR or on a component. @return: the string value of the X-CALENDARSEREVR-MASK-UID property, or None """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: return component.getMaskUID() else: try: # Find the primary subcomponent return self.propertyValue("X-CALENDARSERVER-MASK-UID") except InvalidICalendarDataError: pass return None def getExtendedFreeBusy(self): """ Get the X-CALENDARSERVER-EXTENDED-FREEBUSY value. Works on either a VCALENDAR or on a component. @return: the string value of the X-CALENDARSERVER-EXTENDED-FREEBUSY property, or None """ # Extract appropriate sub-component if this is a VCALENDAR if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() not in ignoredComponents: return component.getExtendedFreeBusy() else: try: # Find the primary subcomponent return self.propertyValue("X-CALENDARSERVER-EXTENDED-FREEBUSY") except InvalidICalendarDataError: pass return None def setParameterToValueForPropertyWithValue(self, paramname, paramvalue, propname, propvalue): """ Add or change the parameter to the specified value on the property having the specified value. @param paramname: the parameter name @type paramname: C{str} @param paramvalue: the parameter value to set @type paramvalue: C{str} @param propname: the property name @type propname: C{str} @param propvalue: the property value to test @type propvalue: C{str} or C{None} """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue for property in component.properties(propname): if propvalue is None or property.value() == propvalue: property.setParameter(paramname, paramvalue) def hasPropertyInAnyComponent(self, properties): """ Test for the existence of one or more properties in any component. @param properties: property name(s) to test for @type properties: C{list}, C{tuple} or C{str} """ if isinstance(properties, str): properties = (properties,) for property in properties: if self.hasProperty(property): return True for component in self.subcomponents(): if component.hasPropertyInAnyComponent(properties): return True return False def getFirstPropertyInAnyComponent(self, properties): """ Get the first of any set of properties in any component. @param properties: property name(s) to test for @type properties: C{list}, C{tuple} or C{str} """ if isinstance(properties, str): properties = (properties,) for property in properties: props = tuple(self.properties(property)) if props: return props[0] for component in self.subcomponents(): prop = component.getFirstPropertyInAnyComponent(properties) if prop: return prop return None def getAllPropertiesInAnyComponent(self, properties, depth=2): """ Get the all of any set of properties in any component down to a specified depth. @param properties: property name(s) to test for @type properties: C{list}, C{tuple} or C{str} @param depth: how deep to go in looking at sub-components: 0: do not go into sub-components, 1: go into one level of sub-components, 2: two levels (which is effectively all the levels supported in iCalendar) @type depth: int """ results = [] if isinstance(properties, str): properties = (properties,) for property in properties: props = tuple(self.properties(property)) if props: results.extend(props) if depth > 0: for component in self.subcomponents(): results.extend(component.getAllPropertiesInAnyComponent(properties, depth - 1)) return results def hasPropertyValueInAllComponents(self, property): """ Test for the existence of a property with a specific value in any sub-component. @param property: property to test for @type property: L{Property} """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue found = component.getProperty(property.name()) if not found or found.value() != property.value(): return False return True def addPropertyToAllComponents(self, property): """ Add a property to all top-level components except VTIMEZONE. @param property: the property to add @type property: L{Property} """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.addProperty(property) def replacePropertyInAllComponents(self, property): """ Replace a property in all components. @param property: the L{Property} to replace in this component. """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.replaceProperty(property) def hasPropertyWithParameterMatch(self, propname, param_name, param_value, param_value_is_default=False): """ See if property whose name, and parameter name, value match in any components. @param property: the L{Property} to replace in this component. @param param_name: the C{str} of parameter name to match. @param param_value: the C{str} of parameter value to match, if C{None} then just match on the presence of the parameter name. @param param_value_is_default: C{bool} to indicate whether absence of the named parameter also implies a match @return: C{True} if matching property found, C{False} if not @rtype: C{bool} """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue if component.hasPropertyWithParameterMatch(propname, param_name, param_value, param_value_is_default): return True else: for oldprop in tuple(self.properties(propname)): pvalue = oldprop.parameterValue(param_name) if pvalue is None and param_value_is_default or pvalue == param_value or param_value is None: return True return False def replaceAllPropertiesWithParameterMatch(self, property, param_name, param_value, param_value_is_default=False): """ Replace a property whose name, and parameter name, value match in all components. @param property: the L{Property} to replace in this component. @param param_name: the C{str} of parameter name to match. @param param_value: the C{str} of parameter value to match. @param param_value_is_default: C{bool} to indicate whether absence of the named parameter also implies a match """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.replaceAllPropertiesWithParameterMatch(property, param_name, param_value, param_value_is_default) else: for oldprop in tuple(self.properties(property.name())): pvalue = oldprop.parameterValue(param_name) if pvalue is None and param_value_is_default or pvalue == param_value: self.removeProperty(oldprop) self.addProperty(property) def removeAllPropertiesWithParameterMatch(self, propname, param_name, param_value, param_value_is_default=False): """ Remove all properties whose name, and parameter name, value match in all components. """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.removeAllPropertiesWithParameterMatch(propname, param_name, param_value, param_value_is_default) else: for oldprop in tuple(self.properties(propname)): pvalue = oldprop.parameterValue(param_name) if pvalue is None and param_value_is_default or pvalue == param_value: self.removeProperty(oldprop) def transferProperties(self, from_calendar, properties): """ Transfer specified properties from old calendar into all components of this calendar, synthesizing any for new overridden instances. @param from_calendar: the old calendar to copy from @type from_calendar: L{Component} @param properties: the property names to copy over @type properties: C{tuple} or C{list} """ assert from_calendar.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.transferProperties(from_calendar, properties) else: # Is there a matching component rid = self.getRecurrenceIDUTC() matched = from_calendar.overriddenComponent(rid) # If no match found, we are processing a new overridden instance so copy from the original master if not matched: matched = from_calendar.masterComponent() if matched: for propname in properties: for prop in matched.properties(propname): self.addProperty(prop) def attendeesView(self, attendees, onlyScheduleAgentServer=False): """ Filter out any components that all attendees are not present in. Use EXDATEs on the master to account for changes. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) # Modify any components that reference the attendee, make note of the ones that don't remove_components = [] master_component = None removed_master = False for component in self.subcomponents(): if component.name() in ignoredComponents: continue found_all_attendees = True for attendee in attendees: foundAttendee = component.getAttendeeProperty((attendee,)) if foundAttendee is None: found_all_attendees = False break if onlyScheduleAgentServer: if foundAttendee.hasParameter("SCHEDULE-AGENT"): if foundAttendee.parameterValue("SCHEDULE-AGENT") != "SERVER": found_all_attendees = False break if not found_all_attendees: remove_components.append(component) if component.getRecurrenceIDUTC() is None: master_component = component if not found_all_attendees: removed_master = True # Now remove the unwanted components - but we may need to EXDATE the master exdates = [] for component in remove_components: rid = component.getRecurrenceIDUTC() if rid is not None: exdates.append(rid) self.removeComponent(component) if not removed_master and master_component is not None: for exdate in exdates: master_component.addProperty(Property("EXDATE", [exdate, ])) def filterComponents(self, rids): # If master is in rids do nothing if not rids or None in rids: return True assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) # Remove components not in the list components = tuple(self.subcomponents()) remaining = len(components) for component in components: if component.name() in ignoredComponents: remaining -= 1 continue rid = component.getRecurrenceIDUTC() if rid not in rids: self.removeComponent(component) remaining -= 1 return remaining != 0 def removeAllButOneAttendee(self, attendee): """ Remove all ATTENDEE properties except for the one specified. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) for component in self.subcomponents(): if component.name() in ignoredComponents: continue [component.removeProperty(p) for p in tuple(component.properties(component.recipientPropertyName())) if p.value().lower() != attendee.lower()] def removeAllButTheseAttendees(self, attendees): """ Remove all ATTENDEE properties except for the ones specified. """ assert self.name() == "VCALENDAR", "Not a calendar: {0!r}".format(self,) attendees = set([attendee.lower() for attendee in attendees]) for component in self.subcomponents(): if component.name() in ignoredComponents: continue [component.removeProperty(p) for p in tuple(component.properties(component.recipientPropertyName())) if p.value().lower() not in attendees] def hasAlarm(self): """ Test whether the component has a VALARM as an immediate sub-component. """ assert self.name().upper() in ("VEVENT", "VTODO",), "Not a VEVENT or VTODO: {0!r}".format(self,) for component in self.subcomponents(): if component.name().upper() == "VALARM": return True return False def addAlarms(self, alarm, ignoreActionNone=True): """ Add an alarm to any VEVENT or VTODO subcomponents that do not already have any. @param alarm: the text for a VALARM component @type alarm: C{str} @param ignoreActionNone: whether or not to skip ACTION:NONE alarms @type ignoreActionNone: C{bool} @return: indicate whether a change was made @rtype: C{bool} """ # Create a fake component for the alarm text caldata = """BEGIN:VCALENDAR VERSION:2.0 PRODID:-//CALENDARSERVER.ORG//NONSGML Version 1//EN BEGIN:VEVENT UID:bogus DTSTART:20110427T000000Z DURATION:PT1H DTSTAMP:20110427T000000Z SUMMARY:bogus {0}END:VEVENT END:VCALENDAR """.replace("\n", "\r\n").format(alarm,) try: calendar = Component.fromString(caldata) if calendar is None: return False except ValueError: return False try: valarm = tuple(tuple(calendar.subcomponents())[0].subcomponents())[0] except IndexError: return False # Need to add property to indicate this was added by the server valarm.addProperty(Property("X-APPLE-DEFAULT-ALARM", "TRUE")) # ACTION:NONE not added changed = False action = valarm.propertyValue("ACTION") if not ignoreActionNone or action and action.upper() != "NONE": for component in self.subcomponents(): if component.name().upper() not in ("VEVENT", "VTODO",): continue if component.hasAlarm(): continue component.addComponent(valarm.duplicate()) changed = True return changed def removeAlarms(self): """ Remove all Alarms components """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.removeAlarms() else: for component in tuple(self.subcomponents()): if component.name() == "VALARM": self.removeComponent(component) def hasDuplicateAlarms(self, doFix=False): """ Test and optionally remove alarms that have the same ACTION and TRIGGER values in the same component. """ changed = False if self.name() in ("VCALENDAR", PERUSER_COMPONENT,): for component in self.subcomponents(): if component.name() in ("VTIMEZONE",): continue changed = component.hasDuplicateAlarms(doFix) or changed else: action_trigger = set() for component in tuple(self.subcomponents()): if component.name() == "VALARM": item = (component.propertyValue("ACTION"), component.propertyValue("TRIGGER"),) if item in action_trigger: if doFix: self.removeComponent(component) changed = True else: action_trigger.add(item) return changed def filterProperties(self, remove=None, keep=None, do_subcomponents=True): """ Remove all properties that do not match the provided set. """ if do_subcomponents: for component in self.subcomponents(): component.filterProperties(remove, keep, do_subcomponents=False) else: if self.name() in ignoredComponents: return for p in tuple(self.properties()): if (keep and p.name() not in keep) or (remove and p.name() in remove): self.removeProperty(p) def removeXComponents(self, keep_components=()): """ Remove all X- components except the specified ones """ for component in tuple(self.subcomponents()): if component.name().startswith("X-") and component.name() not in keep_components: self.removeComponent(component) def removeXProperties(self, keep_properties=(), remove_x_parameters=True, do_subcomponents=True): """ Remove all X- properties except the specified ones """ if do_subcomponents and self.name() == "VCALENDAR": for component in self.subcomponents(): component.removeXProperties(keep_properties, remove_x_parameters, do_subcomponents=False) else: if self.name() in ignoredComponents: return for p in tuple(self.properties()): xpname = p.name().startswith("X-") if xpname and p.name() not in keep_properties: self.removeProperty(p) elif not xpname and remove_x_parameters: for paramname in p.parameterNames(): if paramname.startswith("X-"): p.removeParameter(paramname) def removePropertyParameters(self, property, params): """ Remove all specified property parameters """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.removePropertyParameters(property, params) else: props = self.properties(property) for prop in props: for param in params: prop.removeParameter(param) def removePropertyParametersByValue(self, property, paramvalues): """ Remove all specified property parameters """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.removePropertyParametersByValue(property, paramvalues) else: props = self.properties(property) for prop in props: for param, value in paramvalues: prop.removeParameterValue(param, value) def getITIPInfo(self): """ Get property value details needed to synchronize iTIP components. @return: C{tuple} of (uid, seq, dtstamp, r-id) some of which may be C{None} if property does not exist """ try: # Extract items from component uid = self.propertyValue("UID") seq = self.propertyValue("SEQUENCE") if seq: seq = int(seq) dtstamp = self.propertyValue("DTSTAMP") rid = self.propertyValue("RECURRENCE-ID") except ValueError: return (None, None, None, None) return (uid, seq, dtstamp, rid) @staticmethod def compareComponentsForITIP(component1, component2, use_dtstamp=True): """ Compare synchronization information for two components to see if they match according to iTIP. @param component1: first component to check. @type component1: L{Component} @param component2: second component to check. @type component2: L{Component} @param use_dtstamp: whether DTSTAMP is used in addition to SEQUENCE. @type component2: C{bool} @return: 0, 1, -1 as per compareSyncInfo. """ info1 = (None,) + Component.getITIPInfo(component1) info2 = (None,) + Component.getITIPInfo(component2) return Component.compareITIPInfo(info1, info2, use_dtstamp) @staticmethod def compareITIPInfo(info1, info2, use_dtstamp=True): """ Compare two synchronization information records. @param info1: a C{tuple} as returned by L{getSyncInfo}. @param info2: a C{tuple} as returned by L{getSyncInfo}. @return: 1 if info1 > info2, 0 if info1 == info2, -1 if info1 < info2 """ _ignore_name1, uid1, seq1, dtstamp1, _ignore_rid1 = info1 _ignore_name2, uid2, seq2, dtstamp2, _ignore_rid2 = info2 # UIDs MUST match assert uid1 == uid2 # Look for sequence if (seq1 is not None) and (seq2 is not None): if seq1 > seq2: return 1 if seq1 < seq2: return -1 elif (seq1 is not None) and (seq2 is None): return 1 elif (seq1 is None) and (seq2 is not None): return -1 # Look for DTSTAMP if use_dtstamp: if (dtstamp1 is not None) and (dtstamp2 is not None): if dtstamp1 > dtstamp2: return 1 if dtstamp1 < dtstamp2: return -1 elif (dtstamp1 is not None) and (dtstamp2 is None): return 1 elif (dtstamp1 is None) and (dtstamp2 is not None): return -1 return 0 def needsiTIPSequenceChange(self, oldcalendar): """ Compare this calendar with the old one and indicate whether the current one has SEQUENCE that is always greater than the old. """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue oldcomponent = oldcalendar.overriddenComponent(component.getRecurrenceIDUTC()) if oldcomponent is None: oldcomponent = oldcalendar.masterComponent() if oldcomponent is None: continue newseq = component.propertyValue("SEQUENCE") if newseq is None: newseq = 0 oldseq = oldcomponent.propertyValue("SEQUENCE") if oldseq is None: oldseq = 0 if newseq <= oldseq: return True return False def bumpiTIPInfo(self, oldcalendar=None, doSequence=False): """ Change DTSTAMP and optionally SEQUENCE on all components. """ if doSequence: def maxSequence(calendar): seqs = calendar.getAllPropertiesInAnyComponent("SEQUENCE", depth=1) return max(seqs, key=lambda x: x.value()).value() if seqs else 0 # Determine value to bump to from old calendar (if exists) or self newseq = maxSequence(oldcalendar if oldcalendar is not None else self) + 1 # Bump all components self.replacePropertyInAllComponents(Property("SEQUENCE", newseq)) self.replacePropertyInAllComponents(Property("DTSTAMP", DateTime.getNowUTC())) def sequenceInSync(self, oldcalendar): """ Make sure SEQUENCE does not decrease in any components. """ def maxSequence(calendar): seqs = calendar.getAllPropertiesInAnyComponent("SEQUENCE", depth=1) return max(seqs, key=lambda x: x.value()).value() if seqs else 0 def minSequence(calendar): seqs = calendar.getAllPropertiesInAnyComponent("SEQUENCE", depth=1) return min(seqs, key=lambda x: x.value()).value() if seqs else 0 # Determine value to bump to from old calendar (if exists) or self oldseq = maxSequence(oldcalendar) currentseq = minSequence(self) # Sync all components if oldseq and currentseq < oldseq: self.replacePropertyInAllComponents(Property("SEQUENCE", oldseq)) def normalizeAll(self): # Normalize all properties for prop in tuple(self.properties()): result = normalizeProps.get(prop.name()) if result: default_value, default_params = result else: # Assume default VALUE is TEXT default_value = None default_params = {"VALUE": "TEXT"} # Remove any default parameters for name in prop.parameterNames(): value = prop.parameterValue(name) if value == default_params.get(name): prop.removeParameter(name) # If there are no parameters, remove the property if it has the default value if len(prop.parameterNames()) == 0: if default_value is not None and prop.value() == default_value: self.removeProperty(prop) continue # Otherwise look for value normalization normalize_function = normalizePropsValue.get(prop.name()) if normalize_function: prop.setValue(normalize_function(prop.value())) # Do datetime/rrule normalization self.normalizeDateTimes() # Do to all sub-components too for component in self.subcomponents(): component.normalizeAll() def normalizeDateTimes(self): """ Normalize various datetime properties into UTC and handle DTEND/DURATION variants in such a way that we can compare objects with slight differences. Also normalize the RRULE value parts. Strictly speaking we should not need to do this as clients should not be messing with these properties - i.e. they should round trip them. Unfortunately some do... """ # TODO: what about VJOURNAL and VTODO? if self.name() == "VEVENT": # Basic time properties dtstart = self.getProperty("DTSTART") dtend = self.getProperty("DTEND") duration = self.getProperty("DURATION") timeRange = Period( start=dtstart.value(), end=dtend.value() if dtend is not None else None, duration=duration.value() if duration is not None else None, ) # Have to fake the TZID value here when we convert date-times to UTC # as we need to know what the original one was if dtstart.hasParameter("TZID"): dtstart.setParameter("_TZID", dtstart.parameterValue("TZID")) dtstart.removeParameter("TZID") dtstart.value().adjustToUTC() if dtend is not None: if dtend.hasParameter("TZID"): dtend.setParameter("_TZID", dtend.parameterValue("TZID")) dtend.removeParameter("TZID") dtend.value().adjustToUTC() elif duration is not None: self.removeProperty(duration) self.addProperty(Property("DTEND", timeRange.getEnd().duplicateAsUTC())) rdates = self.properties("RDATE") for rdate in rdates: if rdate.hasParameter("TZID"): rdate.setParameter("_TZID", rdate.parameterValue("TZID")) rdate.removeParameter("TZID") for value in rdate.value(): value.getValue().adjustToUTC() exdates = self.properties("EXDATE") for exdate in exdates: if exdate.hasParameter("TZID"): exdate.setParameter("_TZID", exdate.parameterValue("TZID")) exdate.removeParameter("TZID") for value in exdate.value(): value.getValue().adjustToUTC() rid = self.getProperty("RECURRENCE-ID") if rid is not None: rid.removeParameter("TZID") rid.setValue(rid.value().duplicateAsUTC()) # Recurrence rules - we need to normalize the order of the value parts # for rrule in self._pycalendar.getRecurrenceSet().getRules(): # indexedTokens = {} # indexedTokens.update([valuePart.split("=") for valuePart in rrule.value().split(";")]) # sortedValue = ";".join(["%s=%s" % (key, value,) for key, value in sorted(indexedTokens.iteritems(), key=lambda x:x[0])]) # rrule.setValue(sortedValue) def normalizePropertyValueLists(self, propname): """ Convert properties that have a list of values into single properties, to make it easier to do comparisons between two ical objects. """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.normalizePropertyValueLists(propname) else: for prop in tuple(self.properties(propname)): if type(prop.value()) is list and len(prop.value()) > 1: self.removeProperty(prop) for value in prop.value(): self.addProperty(Property(propname, [value.getValue(), ])) def normalizeAttachments(self): """ Remove any ATTACH properties that relate to a dropbox. """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ignoredComponents: continue component.normalizeAttachments() else: dropboxPrefix = self.propertyValue("X-APPLE-DROPBOX") if dropboxPrefix is None: return for attachment in tuple(self.properties("ATTACH")): valueType = attachment.parameterValue("VALUE") if valueType in (None, "URI"): dataValue = attachment.value() if dataValue.find(dropboxPrefix) != -1: self.removeProperty(attachment) @inlineCallbacks def normalizeCalendarUserAddresses( self, lookupFunction, recordFunction, toCanonical=True ): """ Do the ORGANIZER/ATTENDEE property normalization. @param lookupFunction: function returning full name, guid, CUAs for a given CUA @type lookupFunction: L{Function} @param recordFunction: function taking a CUA and returning a record @type recordFunction: L{Function} @param toCanonical: whether to convert to the canonical CUA form (True) or to the mailto: form (False) @type toCanonical: L{bool} """ for component in self.subcomponents(): if component.name() in ignoredComponents: continue for prop in itertools.chain( component.properties("ORGANIZER"), component.properties("ATTENDEE"), component.properties("VOTER") ): # Check that we can lookup this calendar user address - if not # we cannot do anything with it cuaddr = normalizeCUAddr(prop.value()) name, uid, cutype, cuaddrs = yield lookupFunction(cuaddr, recordFunction, config) if uid is None: continue # Get any EMAIL parameter oldemail = prop.parameterValue("EMAIL") if oldemail: oldemail = "mailto:{0}".format(oldemail,) # Get any CN parameter oldCN = prop.parameterValue("CN") if toCanonical: # Always re-write value to urn:x-uid prop.setValue("urn:x-uid:{uid}".format(uid=uid)) # If it is already a non-x-uid address leave it be elif (cuaddr.startswith("urn:x-uid:") or cuaddr.startswith("urn:uuid:")): if oldemail: # Use the EMAIL parameter if it exists newaddr = oldemail else: # Pick the first mailto, # or failing that the first path one, # or failing that the first http one, # or failing that the first one first_mailto = None first_path = None first_http = None first = None for addr in cuaddrs: if addr.startswith("mailto:"): first_mailto = addr break elif addr.startswith("/"): if not first_path: first_path = addr elif addr.startswith("http:"): if not first_http: first_http = addr elif not first: first = addr if first_mailto: newaddr = first_mailto elif first_path: newaddr = first_path elif first_http: newaddr = first_http elif first: newaddr = first else: newaddr = None # Make the change if newaddr: prop.setValue(newaddr) # Re-write the CN parameter if name: if name != oldCN: prop.setParameter("CN", name) # Also adjust any previously matching location property if cutype == "ROOM": location = component.getProperty("LOCATION") if location is not None: if location.value() == oldCN: location.setValue(name) else: prop.removeParameter("CN") # Re-write the EMAIL if its value no longer matches if oldemail and oldemail not in cuaddrs or oldemail is None and toCanonical: if cuaddr.startswith("mailto:") and cuaddr in cuaddrs: email = cuaddr[7:] else: for addr in cuaddrs: if addr.startswith("mailto:"): email = addr[7:] break else: email = None if email: prop.setParameter("EMAIL", email) else: prop.removeParameter("EMAIL") if cutype != prop.parameterValue("CUTYPE"): if cutype and cutype != "INDIVIDUAL": # For groups we need to change the CUTYPE exposed to clients when we have server-managed # group attendee expansion because some clients seem to spontaneous remove CUTYPE=GROUP # for no obvious reason when the event is changed. We can't have that happen for the # server-managed groups so using a different CUTYPE seems to work around that. if config.GroupAttendees.Enabled and cutype == "GROUP": cutype = "X-SERVER-GROUP" prop.setParameter("CUTYPE", cutype) else: prop.removeParameter("CUTYPE") # For VPOLL also do immediate children if component.name() == "VPOLL": yield component.normalizeCalendarUserAddresses(lookupFunction, recordFunction, toCanonical) def _reconcileGroupAttendee(self, groupCUA, memberAtttendeeProps): changed = False for component in self.subcomponents(): if component.name() in ignoredComponents: continue oldAttendeeProps = tuple(component.properties("ATTENDEE")) oldAttendeeCUAs = set([attendeeProp.value() for attendeeProp in oldAttendeeProps]) # add new member attendees memberCUAs = set() for newAttendeeProp in memberAtttendeeProps: memberCUA = newAttendeeProp.value() if memberCUA not in oldAttendeeCUAs: component.addProperty(newAttendeeProp) changed = True memberCUAs.add(memberCUA) # remove attendee or update MEMBER attribute for non-primary attendees in this group, for attendeeProp in oldAttendeeProps: if attendeeProp.hasParameter("MEMBER"): parameterValues = tuple(attendeeProp.parameterValues("MEMBER")) if groupCUA in parameterValues: if attendeeProp.value() not in memberCUAs: attendeeProp.removeParameterValue("MEMBER", groupCUA) if not attendeeProp.parameterValues("MEMBER"): component.removeProperty(attendeeProp) changed = True else: if attendeeProp.value() in memberCUAs: attendeeProp.setParameter("MEMBER", parameterValues + (groupCUA,)) changed = True return changed def reconcileGroupAttendees(self, groupCUAToAttendeeMemberPropMap): changed = False allMemberCUAs = set() nonemptyGroupCUAs = set() for groupCUA, memberAttendeeProps in groupCUAToAttendeeMemberPropMap.iteritems(): changed |= self._reconcileGroupAttendee(groupCUA, memberAttendeeProps) allMemberCUAs |= set([memberAttendeeProp.value() for memberAttendeeProp in memberAttendeeProps]) if memberAttendeeProps: nonemptyGroupCUAs.add(groupCUA) # remove orphans for component in self.subcomponents(): if component.name() in ignoredComponents: continue for attendeeProp in tuple(component.properties("ATTENDEE")): if attendeeProp.hasParameter("MEMBER"): attendeeCUA = attendeeProp.value() if attendeeCUA in allMemberCUAs: # remove orphan member values parameterValues = tuple(attendeeProp.parameterValues("MEMBER")) for orphanGroupCUA in set(parameterValues) - nonemptyGroupCUAs: attendeeProp.removeParameterValue("MEMBER", orphanGroupCUA) if not attendeeProp.parameterValues("MEMBER"): component.removeProperty(attendeeProp) changed = True else: # remove orphaned member property component.removeProperty(attendeeProp) changed = True return changed def allPerUserUIDs(self): results = set() for component in self.subcomponents(): if component.name() == PERUSER_COMPONENT: results.add(component.propertyValue(PERUSER_UID)) return results def perUserData(self, rid): # We will create a cache of all user/rid/transparency/adjusted_start/adjusted_end values as we will likely # be calling this a lot if not hasattr(self, "_perUserData"): self._perUserData = {} # Do per-user data for component in self.subcomponents(): if component.name() == PERUSER_COMPONENT: uid = component.propertyValue(PERUSER_UID) for subcomponent in component.subcomponents(): if subcomponent.name() == PERINSTANCE_COMPONENT: instancerid = subcomponent.propertyValue("RECURRENCE-ID") transp = subcomponent.propertyValue("TRANSP") == "TRANSPARENT" adjusted_start = subcomponent.propertyValue("X-APPLE-TRAVEL-DURATION") if adjusted_start is None: adjusted_start = subcomponent.propertyValue("X-APPLE-TRAVEL-START") adjusted_end = subcomponent.propertyValue("X-APPLE-TRAVEL-RETURN-DURATION") if adjusted_end is None: adjusted_end = subcomponent.propertyValue("X-APPLE-TRAVEL-RETURN") self._perUserData.setdefault(uid, {})[instancerid] = (transp, adjusted_start, adjusted_end,) elif component.name() not in ignoredComponents: instancerid = component.propertyValue("RECURRENCE-ID") transp = component.propertyValue("TRANSP") == "TRANSPARENT" self._perUserData.setdefault("", {})[instancerid] = (transp, None, None,) # Now lookup in cache results = [] for uid, cachedRids in sorted(self._perUserData.items(), key=lambda x: x[0]): lookupRid = rid if lookupRid not in cachedRids: lookupRid = None if lookupRid in cachedRids: results.append((uid, cachedRids[lookupRid],)) else: results.append((uid, (False, None, None,))) return tuple(results) def hasInstancesAfter(self, limit): """ Determine whether an event exists completely prior to a given moment. @param limit: the moment to compare against. @type limit: L{DateTime} @return: a C{bool}, True if the event has any instances occurring after limit, False otherwise. """ instanceList = self.expandTimeRanges(limit) if instanceList.limit is not None: # There are instances after the limit return True # All instances begin prior to limit, but now check their end times to # see if they extend beyond limit for instance in instanceList.instances.itervalues(): if instance.end > limit: return True # Exists completely prior to limit return False def hasDuplicatePrivateComments(self, doFix=False): """ Test and optionally remove "X-CALENDARSERVER-ATTENDEE-COMMENT" properties that have the same "X-CALENDARSERVER-ATTENDEE-REF" parameter values in the same component. @return: C{True} if there are duplicates that were not fixed. """ if self.name() == "VCALENDAR": for component in self.subcomponents(): if component.name() in ("VTIMEZONE",): continue if component.hasDuplicatePrivateComments(doFix): return True else: attendee_refs = set() for prop in tuple(self.properties("X-CALENDARSERVER-ATTENDEE-COMMENT")): ref = prop.parameterValue("X-CALENDARSERVER-ATTENDEE-REF") if ref in attendee_refs: if doFix: self.removeProperty(prop) else: return True attendee_refs.add(ref) return False # # # Timezones # # def tzexpand(tzdata, start, end): """ Expand a timezone to get onset/utc-offset observance tuples within the specified time range. @param tzdata: the iCalendar data containing a VTIMEZONE. @type tzdata: C{str} @param start: date for the start of the expansion. @type start: C{date} @param end: date for the end of the expansion. @type end: C{date} @return: a C{list} of tuples of (C{datetime}, C{str}) """ icalobj = Component.fromString(tzdata) tzcomp = None for comp in icalobj.subcomponents(): if comp.name() == "VTIMEZONE": tzcomp = comp break else: raise InvalidICalendarDataError("No VTIMEZONE component in {0}".format(tzdata,)) tzexpanded = tzcomp._pycalendar.expandAll(start, end) results = [] # Always need to ensure the start appears in the result start.setDateOnly(False) if tzexpanded: if start != tzexpanded[0][0]: results.append((str(start), UTCOffsetValue(tzexpanded[0][2]).getText(),)) else: results.append((str(start), UTCOffsetValue(tzcomp._pycalendar.getTimezoneOffsetSeconds(start)).getText(),)) for tzstart, _ignore_utctzstart, _ignore_tzoffsetfrom, tzoffsetto in tzexpanded: results.append(( tzstart.getText(), UTCOffsetValue(tzoffsetto).getText(), )) return results def tzexpandlocal(tzdata, start, end): """ Expand a timezone to get onset(local)/utc-offset-from/utc-offset-to/name observance tuples within the specified time range. @param tzdata: the iCalendar data containing a VTIMEZONE. @type tzdata: L{Calendar} @param start: date for the start of the expansion. @type start: C{date} @param end: date for the end of the expansion. @type end: C{date} @return: a C{list} of tuples """ icalobj = Component(None, pycalendar=tzdata) tzcomp = None for comp in icalobj.subcomponents(): if comp.name() == "VTIMEZONE": tzcomp = comp break else: raise InvalidICalendarDataError("No VTIMEZONE component in {0}".format(tzdata,)) tzexpanded = tzcomp._pycalendar.expandAll(start, end, with_name=True) results = [] # Always need to ensure the start appears in the result start.setDateOnly(False) if tzexpanded: if start != tzexpanded[0][0]: results.append(( start, tzexpanded[0][2], tzexpanded[0][2], tzexpanded[0][4], )) else: results.append(( start, tzcomp._pycalendar.getTimezoneOffsetSeconds(start), tzcomp._pycalendar.getTimezoneOffsetSeconds(start), tzcomp._pycalendar.getTimezoneDescriptor(start), )) for tzstart, _ignore_utctzstart, tzoffsetfrom, tzoffsetto, name in tzexpanded: results.append(( tzstart, tzoffsetfrom, tzoffsetto, name, )) return results # # # Utilities # # @inlineCallbacks def normalizeCUAddress(cuaddr, lookupFunction, recordFunction, toCanonical=True): # Check that we can lookup this calendar user address - if not # we cannot do anything with it _ignore_name, uid, _ignore_cuType, cuaddrs = (yield lookupFunction(normalizeCUAddr(cuaddr), recordFunction, config)) if toCanonical: # Always re-write value to urn:x-uid if uid: returnValue("urn:x-uid:{0}".format(uid,)) # If it is already a non-x-uid address leave it be elif (cuaddr.startswith("urn:x-uid:") or cuaddr.startswith("urn:uuid:")): # Pick the first mailto, # or failing that the first path one, # or failing that the first http one, # or failing that the first one first_mailto = None first_path = None first_http = None first = None for addr in cuaddrs: if addr.startswith("mailto:"): first_mailto = addr break elif addr.startswith("/"): if not first_path: first_path = addr elif addr.startswith("http:"): if not first_http: first_http = addr elif not first: first = addr if first_mailto: newaddr = first_mailto elif first_path: newaddr = first_path elif first_http: newaddr = first_http elif first: newaddr = first else: newaddr = None # Make the change if newaddr: returnValue(newaddr) returnValue(cuaddr) # # This function is from "Python Cookbook, 2d Ed., by Alex Martelli, Anna # Martelli Ravenscroft, and David Ascher (O'Reilly Media, 2005) 0-596-00797-3." # def merge(*iterables): """ Merge sorted iterables into one sorted iterable. @param iterables: arguments are iterables which yield items in sorted order. @return: an iterable of all items generated by every iterable in C{iterables} in sorted order. """ heap = [] for iterable in iterables: iterator = iter(iterable) for value in iterator: heap.append((value, iterator)) break heapq.heapify(heap) while heap: value, iterator = heap[0] yield value for value in iterator: heapq.heapreplace(heap, (value, iterator)) break else: heapq.heappop(heap) def normalize_iCalStr(icalstr): """ Normalize a string representation of ical data for easy test comparison. """ icalstr = str(icalstr).replace("\r\n ", "") icalstr = icalstr.replace("\n ", "") lines = [line for line in icalstr.splitlines() if not line.startswith("DTSTAMP")] for ctr, line in enumerate(lines[:]): pos = line.find(";X-CALENDARSERVER-DTSTAMP=") if pos != -1: lines[ctr] = line[:pos] + line[pos + len(";X-CALENDARSERVER-DTSTAMP=") + 16:] icalstr = "\r\n".join(lines) return icalstr + "\r\n" def diff_iCalStrs(icalstr1, icalstr2): icalstr1 = normalize_iCalStr(icalstr1).splitlines() icalstr2 = normalize_iCalStr(icalstr2).splitlines() return "\n".join(unified_diff(icalstr1, icalstr2))
A form of potential energy is set up between electric charges analagous to the case when two masses are separated. The easiest way to understand electrical energy is to consider a capacitor. In a capacitor a potential difference is set up (by a battery) between two conducting plates. This voltage difference V encourages charges to move from one plate to the other. Since like charges repel, work needs to be done to force more of them onto the same plate. The work required dW to move an increment dq of charge is given by¹ dW = dqV In a capacitor, the total amount of charge that may be moved in this way is proportional to the potential difference V. The constant of proportionality in this case is known as the capacitance q = CV Substituting this expression for q into the first expression and integrating over q from zero to Q, W = ∫ (q/C) dq W = Q²/2C = (C/2)V² The electic field E is given by the gradient of the potential difference. Since this changes linearly over a distance d from one conductor to the other we have E=Vd. Furthermore the capacitance may be written C = ε A/d where A is the area of the conducting plates. Thus, the work done is given by, W= (ε A/2d)(Ed)² W= (1/2)εE² Ad Note that the product of the area and the gap between the plates is the volume of the capacitor. Dividing the total energy by the volume yields the electrical energy density uE uE = (1/2)εE² In the case of a vacuum , the permittivity ε is that of free space εo . When a dielectric is used between the plates more electrial energy is stored in polarizing See also Poynting vector where the electromagnetic energy density is described and magnetic field energy density. ¹ The work done in moving a mass through a gravitational potential is similarly described.
A few days ago I wrote about the ENCODE project and the new recognition that, at minimum, 80 percent of the human genome consists of functional DNA elements. Despite some skeptics’ complaints that the media, creationists, and intelligent design adherents have misconstrued the ENCODE report, the project’s results stand. The human genome is not a vast wasteland of junk, but appears to be an elegantly designed system. The ENCODE team approached the genome looking for functional elements that play a role in controlling gene expression. But DNA elements can be functional in other ways. For example, some researchers have suggested that the amount of DNA in the genome plays a role controlling the size of the nucleus; that is, DNA also plays a nucleoskeletal role. To learn more about the nucleoskeletal hypothesis, click on the following links: - “More on the Nucleoskeletal Hypothesis” - “Dinosaur Genome Size Estimates: Lagerstatten of Design” - “‘Junk’ DNA Not So Junky”
# Choosing when intitial velocity is zero 1. Aug 19, 2015 ### Mr Davis 97 I have a quick question about when we can define the initial velocity to be zero. For example, in projectile motion problems, even though we know that projectile is originally at rest in a cannon, we assign the initial velocity to be some number that we can work with. However, for problems involving the work-energy theorem, no matter whether the ball has some measurable velocity at the beginning, on earth, we say that the initial velocity is zero in order to conclude that the work done is opposite that of gravity. Therefore, it seems that we pick and choose to define when velocity is zero. Why can we say it's zero in one type of problem (like work-energy theorem problems) but not other problems (liek projectile motion problems)? 2. Aug 19, 2015 ### olivermsun A little clarification would be helpful. In the projectile motion problem, do you mean we give the initial velocity to be whatever velocity the projectile had when it left the barrel of the cannon? On the other hand, what kind of problem are you thinking of with the ball? Can you explain the example in some more detail? 3. Aug 19, 2015 ### SteamKing Staff Emeritus That's not necessarily true. If a projectile is at rest in the cannon's bore before firing, it's going to stay there until the cannon is fired. But, the projectile resting comfortably inside the cannon is not what's interesting about this situation. As the projectile leaves the muzzle of the cannon, it's going to be traveling at some definite velocity, having been accelerated by the expanding gases produced by the burning of the propellant with which the projectile is fired. After the projectile leaves the muzzle, the expanding gases from the propellant are no longer assumed to be accelerating the projectile, hence its instantaneous velocity at that time can be taken as a constant. After that, gravity, air drag, etc., may or may not influence the motion of the projectile, as required by the problem statement. While it is true that some problems have been simplified for teaching purposes, not everything about a given problem is necessarily made up. 4. Aug 19, 2015 ### Staff: Mentor This is not always true in my experience. I have worked many such problems where the initial velocity was not zero.
कुछ ऑनलाइन कंपनियों ने अपने ब्रांड को बढ़ाने के लिए एक कंपनी की स्थापना की थी लेकिन जिस कंपनी से ग्राहकों को सामान भेजा जाता था उसका कुछ और नाम था. इस तरीके से काम करने में अब बदलाव आ सकता है. जयपुर में सक्रीय महिला चोर गैंग,शोरूम से चुराती है कीमती सामान 2. सेक्टर88 से पुलिस ने मुठभेड़ में किया गिरफ्तार मुख्यमंत्री जय राम ठाकुर ने आज सिरमौर ज़िला मुख्यालय नाहन के ऐतिहासिक चौगान में नगर परिषद नाहन की 150वीं वर्षगांठ समारोह के अवसर पर भारी जन समूह को सम्बोधित करते हुए नाहन के लोगों को इस ऐतिहासिक क्षण का हिस्सा बनने के लिए बधाई दी। उन्होंने कहा कि नाहन शहर न केवल... आष्टा। हमेशा अपने बयानों से विवादित रहने वाले पूर्व आप पार्टी के नेता और वकील प्रशांत भूषण का पुतला सोमवार 3 अप्रैल को नगर के राधाकृष्ण मंदिर के पास हिंदू संगठनों ने जलाया। संजय मांजरेकर ने कहा, हमें संजू के भविष्य के बारे में अब ज्यादा बातें नहीं करनी चाहिए। एक खिलाडी की तारीफ होनी चाहिए जिस दिन उसने अच्छा किया हो। हमें उनकी पारी की तारीफ करनी चाहिए लेकिन उनके भविष्य पर बात करना सही नहीं है। वलीदपुर अतरौली, अलीगढ़, उत्तर प्रदेश स्थित एक गाँव</s>
जम्मू-कश्मीर : कुलगाम में आतंकियों ने पांच मजदूरों को गोलियों से भूना कुलगाम में हुए इस हमले में मारे गए पांचों मजदूर गैर कश्मीरी हैं जम्मू-कश्मीर में आतंकियों की बौखलाहट एक बार फिर सामने आई है. कश्मीर के कुलगाम में आतंकी हमला हुआ है जिसमें पांच मजदूरों की मौत हो गई है, जबकि एक घायल है. मारे गए सभी मजदूर कश्मीर से बाहर के हैं. जम्मू-कश्मीर से धारा-३७० हटाए जाने के बाद से घाटी में ये सबसे बड़ा आतंकी हमला है. समाचार एजेंसी एएनआई के मुताबिक जम्मू-कश्मीर पुलिस ने बताया है कि सुरक्षा बलों ने कुलगाम आसपास के इलाकों की घेराबंदी कर ली है और बड़े पैमाने पर तलाशी अभियान चल रहा है. अतिरिक्त सुरक्षा बलों को भी बुलाया गया है. माना जा रहा है कि मारे गए मजदूर पश्चिम बंगाल के थे, जो दिहाड़ी पर काम करते थे. कश्मीर में ये आतंकी हमला ऐसे समय हुआ है जब यूरोपीय संघ के २८ सांसद कश्मीर के दौरे पर हैं. सांसदों के दौरे के कारण घाटी में सुरक्षा काफी कड़ी है. इसके बावजूद आतंकी बौखलाहट में किसी ना किसी वारदात को अंजाम दे रहे हैं. सोमवार को ही अनंतनाग में आतंकियों ने एक ट्रक ड्राइवर को गोली मार दी थी और सोपोर में एक बस स्टैंड में बस के इंतजार में बैठे लोगों को भी निशाना बनाया था. यूरोपीय सांसदों के इस दौरे के बीच ही श्रीनगर और दक्षिण कश्मीर के कुछ इलाकों में पत्थरबाजी की घटनाएं भी सामने आईं हैं.
Of the 160 million people living in the Horn of Africa, 70 million live in areas prone to extreme food shortages. 60% of the land is home to 22 million pastoralists. More than 40% of the population in the region is undernourished due to food insecurity and inadequate livelihoods. During the 2010-2011 droughts malnutrition was as high as 30%. Over the past 30 years, frequent famine - every five to seven years - and frequent conflicts have made this region one of the most food-insecure in the world. Urban poverty was the third greatest contributor to food insecurity in the region. Livelihoods and food security, therefore, are priority aspects of peace and development in this region. This has been difficult to achieve because of natural and man-made disasters including, human migration and conflict. The Horn of Africa has a long history of conflicts at the local community level and across borders between communities of different countries. Conflicts disrupt households and their support networks in livelihoods, community social life, food and livestock production, human movement, and commerce; leading to displacement of populations raising vulnerability factors at households adding to exposure to more risks. The context of conflict has taken different dimensions in the Horn region in the last ten years. Apart from internal conflict, piracy, terrorism and political violence have brought new facets to conflict. In addition, the migratory lifestyle of pastoralists has been under threat and instability caused by conflict, change in governance and administration systems, climate change, environmental degradation and natural disasters. Conflict and instability can severely affect socio-cultural, political and power relations. Conflict exacerbates the vulnerability of poor people, displacing them from their homes and depleting their assets. It makes emergency relief operations directed towards IDPs or refugees difficult and dangerous for those involved and has a much more negative impact on long-term development, diverting scarce resources away from development objective into war. In this article, human migration is taken to mean the movement of human populations to settle in other geographic locations, due to any other cause other than intended and voluntary movement. In this regard, movement associated with disaster, conflict, or forced migration will be perceived to constitute human migration. Nomadic movement of communities will not be regarded as part of human migration as it does not in itself constitute an intention for settling. However, where some factors or forces associated with nomadic lifestyles enforce desire to flee or settle away from particular locations to avoid conflict, or attain safety; such migration will be categorized as part of trans-human movement and will be part of human migration. In the Horn of Africa region, a number of factors have led to human migration. These include factors as wide ranging as natural disasters, conflict, poor governance, and lifestyle changes. Effects on livelihoods and food security Conflict and human migration within the Horn of Africa brings with it a myriad of consequences for communities. The key consequences include: - Increased causes of food insecurity for migrant groups, host communities and informal urban households - Increase in new types of vulnerabilities, disadvantage and risks - Reduced coping mechanisms among migrants and host communities - Inadequate living conditions as relates to incomes; access to services and opportunities - Low human and land productivity for lack of productive know-how, inputs and tools - Increased pressure on local resources - land and spaces for shelter, basic services, shared resources, such as water, sanitation, schools, health services and sources of food. - Opportunities for employment, whether formal or informal, are stressed with increased competition. - Over-reliance on humanitarian support – Communities like pastoralists, fisher folk and informal urban households are forced to turn and be depended on humanitarian relief - Barriers of access to land and water, e.g. land allocated to IDPs and refugees aare often not the best for production. - Denied opportunities for production or work – local authorities impose barriers such as forms and levels of taxation, determining types of business or locations where IDPs and refugees can conduct their business. - Resistance by host communities - host communities perceive unfair competition or lack of sufficient market for produce due to competition from refugees and IDPs - Strained relations and transactions between migrants and host communities; or local authorities and informal land settlements It is clear that the problem of livelihoods and food insecurity is complex and multifaceted one in which investments are needed focusing know-how for developing systems, education, health, energy and infrastructure to provide the framework that allows people to broaden livelihoods options that create long-term food security opportunities.
# What 95% RTP truely means.... Discussion in 'Slots Discussion' started by tataru, Apr 22, 2015. Apr 22, 2015 1. ### tataruExperienced Member Occupation: NOT playing slots Location: germany I've read a few posts where it's obvious the posters do not understand what the RTP does. They seem to think that when they deposit 100\$, with a 95% RTP, on average they'd get get back 95\$ - long term. Alas, this is not true at all. If you think this way, it may lead to very bad decision making. Now, how does the RTP work? It may seem to be a small or no difference (to the idea above) at first but trust me, it's a HUGE difference. Every dollar you wager, -long term- you'd get back 95%. Let's take the 100\$ deposit from above. If you wagered 100\$, eg. 100 x 1\$ bet or 200x50\$, you'd get back 95\$. 100\$ * 95% = 95\$ Ok, but now you'want to play a few more spins, at least if you are like me Let's take the amount we have statistically left: 95\$. So, this means you'd take the 95\$ and wager another 95 bets of 1\$. Again, you'd be returned 95%. 95\$ * 95% = 90.25\$ See? After betting close to 200\$ our RTP is just shy above 90%. Of course the RTP per single bet is fixed and is 95% the RTP for your deposit continually decrease.... everytime. What results do you have to expect, if you'd deposit 100\$ and did 500 1\$ bets? (The same holds true for any equivlant amount doing different denomination bets, eg. 250 2\$ bets or 1000 50cent bets) 100\$ * 95% = 95\$ 95\$ * 95% = 90.25\$ 90.25\$ * 95% = 85.74\$ Ok, after betting 285\$ were down to an expected return of 85.74\$ on our 100\$deposit. Let's see what happens if we gamble some more... 85.74\$ * 95% = 81.45\$ 81.45\$ * 95% = 77.38\$ So betting an additional 167\$, totalling 452\$, we're down to an expected return of 77.38\$ for our 100\$ deposit. Obviously our bankroll dwindles, being slowly degisted by the slot. Let's hope he feels satiated I just wanted to clarify to some that the RTP of XX % is often very misunderstood. As you can see, someone who believes a 95% RTP will result in getting 95% of his deposits back is gravely mistaken. If this is still not clear to you, i'd be happy to answer your questions. I think it's very important to understand the meaning of the RTP. I hope this helps 2. Apr 22, 2015 Occupation: Location: Australia I thought RTP is based on the machine life and not a player's lifetime. Take a land based casino, the slot machine is set to 95% via the ROM. Now if a player plays and wins with 200% RTP, eventually the machine will have to adjust itself to get back to 95%. So the next player may be unfortunate. So it's misleading when people say 95%. Who is going to be playing at the same machine, for their entire life, to achieve the 95%? Hence why people complain on here they had 30% RTP, well another player might have had 140% RTP and hence you paid for their winnings. Such is life, someone has to lose for someone else to win. 2 people like this. 4. Apr 22, 2015 5. ### tataruExperienced Member Occupation: NOT playing slots Location: germany My point is, in a nutshell: If you believe deposting amount and RTP have a relationship as explained in my OP you are sorely mistaken and will most likely try to chase "what's yours" even though it's not (statistically speaking). -> You expect too much! 6. Apr 22, 2015 7. ### brianmonMeister Memberwebbymm4 Location: uk The term Return To PlayerS would be more accurate. Since it's more about what the slot will pay, overall to ALL players, than to the individual. Although that doesn't mean EVERY player will get 95%. Just that 95% or all the money from all the players, should in theory, be paid back to the players. Someone could have a 'lifetime' RTP of 70% while another player could have 120% 2 people like this. 8. Apr 22, 2015 9. ### tataruExperienced Member Occupation: NOT playing slots Location: germany My point remains valid though. It's important to realize that one typically bets ones deposit multiple times, thus inreasing the house edge per deposit dollar. This is no spectular insight, to be honest, but I have to recall this to myself every so often. Otherwise I get to feel that a machine "owes" me an we all know that is bad news. 10. Apr 22, 2015 11. ### brianmonMeister Memberwebbymm4 Location: uk But also a winning session doesn't necessarily mean your next session will be a losing one, and a 'personal' rtp on slots can be more flexible than on other games. My lifetime RTP at Betat is currently at just over 100%, it has been as low as 85% and as high as 130%, and that's over a couple of years and 100's of deposits. 12. Apr 22, 2015 13. ### tataruExperienced Member Occupation: NOT playing slots Location: germany Ja, sure, of course things like happen because of the variance. It's the difference between probability and frequency of occurence. I'd like to try my luck today rather than tomorrow, but alas, I'm out of gambling money. If ijust had a 10er Yesterday I played up to 50 bucks on a 10EUR deposit but did not cash out. You all know what happened to my wins 14. Apr 22, 2015 15. ### dunoverUnofficial T&C's EditorStaff MemberCAGPABnononaccredPABnonaccredPABinitmm3webmeister Occupation: International Money Launderer Location: the bus shelter, opposite GCHQ Benhall Could have saved yourself the bother of all that typing! This is a bit more detailed and shows the difference to length gameplay and potential losses just a mere drop in TRTP between games can have... P.S. You are not differentiating between RTP and TRTP. RTP can be fluid and thus leads to variations which obviously means some do far better or worse than your example. Your example is actually based on TRTP. Basically if you bet \$1 on a slot and hit each reel permutation in turn once then stopped, you would receive exactly the stated TRTP back. You are confusing the cash balance effect of RTP with RTP itself - of course the longer you play the bigger proportion of your deposit will be eaten by the 5% house edge, we know that! However, in your example even if the player busts out of his \$100 deposit, he has had 95c back for each dollar he spent during playing, so did get a 95% RTP! Let's say the TRTP was 99.9% - the player keeps playing but will still bust out eventually as the 0.1% house edge eats the \$100 deposit. When he's bust out, his RTP would have been 99.9%. Going by your mistaken logic it would be 0%!! I'm sure one of the reps will put you straight. RTP is NOT worked out by what you have left of your deposit, but by total value of spins you had, minus total amount of wins you had. So, say your \$100 deposit lasted for 1000x\$1 spins, and you had wins of \$900 you would have busted out with a RTP of 90% If you had wins of \$960 you would have lost \$40 and had a RTP of 96%. To suggest we don't understand what RTP does is quite a thing to say - it's obvious that you yourself have got a bit twisted on the subject! 7 people like this. 16. Apr 22, 2015 17. ### Wild ReelsMeister Memberwebmeister Occupation: IT Location: United Kingdom if you have wagered 200\$ and have a balance \$90.25 that means over the course of 200 spins you have lost \$9.75 giving you an overall rtp 190.25/200 x 100 = 95.125% and the slot is behaving as it should do, use rtp as a quick guide to whether its worth playing, there are some nasty online slots in the lower regions of 90%, and some as high as 98% i only use this as a guide of whether its worth playing it, after that its your balance in a session that matters and it hopefully being more than you started with chasing slots to improve rtp is futile, you just wagered more money to improve a figure that if its still below 100% means youve probably lost even more along the way. 18. Apr 22, 2015 19. ### tataruExperienced Member Occupation: NOT playing slots Location: germany You are right of course. I noticed that my post wasn't clear but was too lazy to rewrite SORRY So what's the T in TRTP? 1 person likes this. 20. Apr 22, 2015 21. ### Jono777Meister MemberCAGmm4mm1 Occupation: Self- Employeed Location: Wolverhampton *See posts below, mistake admitted *holds hands up* Last edited: Apr 22, 2015 1 person likes this. 22. Apr 22, 2015 23. ### dunoverUnofficial T&C's EditorStaff MemberCAGPABnononaccredPABnonaccredPABinitmm3webmeister Occupation: International Money Launderer Location: the bus shelter, opposite GCHQ Benhall 'Theoretical'. It's based on the reel permutations occurring once each. Say the number of reel permutations on a slot is 1.5 million on a 5-reel game. These results add up to 1.35 million dollars if you got each result once on a \$1 spin. Therefore you spent \$1.5 million getting \$1.35 million back. Therefore the TRTP is 90%. Of course we never play that amount of spins on any game, normally just a few thousand or tens-of-thousands. Therefore most of us will see 81% or 88.3% or 97.9% or 104.6% i.e. we've all taken a chunk of results from those available. If you're lucky you randomly picked a positive chunk which yielded a profit, or like most of us a crap chunk which means you lose a few hundred playing those few-thousand spins.... Unless you're Rolastan who has played 42.34 billion spins on Dead Or Alive and seems to have a RTP of 300%... 5 people like this. 24. Apr 22, 2015 25. ### tataruExperienced Member Occupation: NOT playing slots Location: germany What you say is opposite idea of T(rue)RTP! I guess there is no common definition of it. 26. Apr 22, 2015 27. ### Wild ReelsMeister Memberwebmeister Occupation: IT Location: United Kingdom Its always been theoretical as far as I'm concerned, it most cases when we are talking about slot rtp we most likely should be saying trtp but we don't out of habit 2 people like this. 28. Apr 22, 2015 29. ### Jono777Meister MemberCAGmm4mm1 Occupation: Self- Employeed Location: Wolverhampton 30. Apr 23, 2015 31. ### nikantwUeber MeisterCAGMMPABaccred Occupation: A bit of this and that Location: EU RTP is (total wins)/(total bets). Has nothing to do with deposits. You lose everything if you play long enough! If you think 95% is low, try to be forced to play with 40%, because it is the only "legal"! But what I want to know is this. What about bet size? Even in the long term, nobody makes the same amount of bets on every bet size. For most people, there would be small bets mostly, and a few big ones. There are monthly reports of RTP. I find it hard to believe that in one month, every month, the number of big bets is big enough to make average RTP come "naturaly". In theory, an RNG decides how often a winning combination comes, regardles of bet size. But if All the loses are with small bets and all the wins with big ones (gamblers paradise ) doesnt it change the RTP? It can happen in a short term, like a month. How do they manage to stay close to 95% every month, even the small casinos? 1 person likes this. 32. Apr 23, 2015 33. ### dunoverUnofficial T&C's EditorStaff MemberCAGPABnononaccredPABnonaccredPABinitmm3webmeister Occupation: International Money Launderer Location: the bus shelter, opposite GCHQ Benhall This is a question I've often pondered. The only way I can see to make this work out is that each player has their own RTP on any particular game at any particular casino. Otherwise, I don't know. 34. Apr 23, 2015 35. ### brianmonMeister Memberwebbymm4 Location: uk Not sure what you mean by that, lol 36. Apr 23, 2015 37. ### dunoverUnofficial T&C's EditorStaff MemberCAGPABnononaccredPABnonaccredPABinitmm3webmeister Occupation: International Money Launderer Location: the bus shelter, opposite GCHQ Benhall I.E. Has a single-use of the game's pool of outcomes via his account - not playing the 'same' game as you or me. 38. Apr 23, 2015 39. ### brianmonMeister Memberwebbymm4 Location: uk You might be needing one of your own helmets here, lol There's supposedly no such thing as a 'pool of wins' and they're supposed to be completely random, fed from one 'master' RNG, or at least that's how the MGS website makes it sound. If there's a pool of wins, does that mean that everyone will eventually get a 5-reel WD? well at least someone put plenty of wildlines in my DOA pool, lol
(Ancient Greek - neuron = nerve ; pteron = wing) Common predators of other insects, including aphids, relatively large wings Insects in this order Common green lacewing © Roger Key The Common Green Lacewing (Chrysoperla carnea) is one of 18 species of green lacewings that live in Britain. Two of these species are actually brownish but all the others are various shades of green. All have long antennae, large golden eyes and two pairs of transparent wings that...
Keep children away! It causes trauma Earthquakes create effects that last for years. It is not necessary to have experienced an earthquake to be struck by an earthquake. Every image seen on screen and on social media can wear down psychologically. Psychologist İrem Naz Kırmızı stated that earthquakes, like any natural disaster, have a devastating effect on human psychology and that children can be adversely affected by this process, saying: “Reduce children’s screen exposure, do not cause secondary trauma, even if there is no trauma in the first place. Do not talk too much about this situation, do not show news images. Not every image and sound is suitable for every age. Noting that earthquakes, like any natural disaster, have a devastating effect on human psychology and that their children may be adversely affected by this process, Psk. “Knowing, recognizing and interpreting these effects makes it easier for us to deal with the psychological chaos it causes. Because people react differently to each event, they will also differ in the process of understanding the natural disaster. While the initial shock effect that will occur at the time of the incident takes place in a quiet and introverted manner, some may face it by crying and whimpering. Stating that children can also be adversely affected by this process, Psk. “Reduce children’s screen exposure, do not cause secondary trauma, even if there is no trauma in the first. Don’t talk to him too much about this situation, don’t show the images. Not every news image and sound about the earthquake is suitable for all ages. You need to be clear and informative, without exaggerating what is available. Listen carefully to their questions about the earthquake and try to answer them without glare. PS. İrem Naz Kırmızı said the following about how to approach people with post-traumatic mood disorders: “Emotions that can arise after a natural disaster can take the form of fear, anger, guilt, helplessness, hopelessness or a frozen image. You may become confused, have difficulty distinguishing between time and space, understand what is happening and remain undecided. You may be physically tense, tired, have trouble sleeping, pain, an irregular heart rhythm, nausea and loss of appetite. You may experience problems in your social life, restlessness, insecurity, anger, difficulty controlling, feeling abandoned, feeling guilty while eating or sleeping, being overly blaming, feeling empty. It should be a priority to meet the physical needs of people we see in such a situation. It is necessary to make the person feel safe after meeting their food, shelter and medical needs. If he wants to tell his story, listening should make him feel understood and not alone. Providing accurate and explanatory clear information about the situation from the right sources, information that we confirm is from reliable sources, must be conveyed without concealing the truth, without confusing it. It is necessary to take the person to the nearest aid agency and help prepare the way for contact with the survivors.” Psk said children show different reactions than adults. Crimea shared the following information: “Children pass on the same reactions through observation that you show in a disaster. How much the child is affected by the earthquake depends on his age. Because children under school age cannot express their feelings and thoughts like adults, they may display physical symptoms such as bed-wetting, thumb sucking, difficulty sleeping at night, staying with you, getting angry and crying frequently, nausea and vomiting. They may fear that they will experience the earthquake again and again.” “Note that your child should be told that the earthquake is a natural occurrence, Psk. “Explain that it is normal to be afraid and to have nightmares, and that it is normal to cry and be sad. There was an earthquake in our country and our house was destroyed, we will not be able to stay in our house for a while, but now the place where we stay is safe and I am with you, I will not leave you. Look, a lot of people on this issue are working to keep us safe and help people. You need to be clear and informative, without exaggerating what is available. Listen carefully to their questions about the earthquake and try to answer them without glare. If he asks for it, get in touch with more and show interest.” Emphasizing that children should not see the earthquake-related images, Psk. “Reduce children’s screen exposure, do not cause secondary trauma, even if there is no trauma in the first. Don’t talk to him too much about this situation, don’t show the images. Not all images and sounds are suitable for all ages. They can’t think like you, remember. After a while, seek help from psychotherapists in the field of trauma, both individually and for children.”
# print.py -- source test pattern for print statements # # This simple program is part of the decompyle test suite. # # decompyle is a Python byte-code decompiler # See http://www.goebel-consult.de/decompyle/ for download and # for further information print 1,2,3,4,5 a = b + 5 print 1,2,3,4,5 print 1,2,3,4,5 print print print 1,2,3,4,5 print
Portuguese scientists discover new mechanism that regulates formation of blood vessels Researchers in one of the external groups of the Instituto Gulbenkian de Ciência (IGC), in Portugal, have discovered a novel mechanism which regulates the process whereby new blood vessels are formed and wounds heal, including chronic wounds, such as those found in diabetic patients and those suffering from morbid obesity. These findings, by Sérgio Dias and his team, are to appear in the new issue of the journal PLoS ONE, and have implications for the development of new therapeutic approaches to healing damaged blood vessels and building new ones. Working at the Centro de Investigação e Patobiologia Molecular of the Portuguese Institute of Oncology Francisco Gentil, in Lisbon, the team showed that the cells that make new blood vessels (called endothelial cells) are stimulated by an intracellular signalling pathway, mediated by the protein Notch. The formation of new blood vessels is a crucial step in wound healing: the newly-formed vessels allow anti-inflammatory proteins to reach the wound site, improve oxygenation of the damaged tissue and carry essential nutrients for the re-structuring of the tissue, that is, the skin. According to Francisco Caiado, a PhD student at the IGC, and first author of this study, "We knew that the endothelial cells are stimulated by cells originating in the bone-marrow, the so-called bone-marrow derived precursor cells. We have now shown that the actual stimulus happens through the Notch protein, found on the bone-marrow derived cells. Upon activation, Notch promotes the adhesion of the precursor cells to the site of the lesion, where they stimulate the endothelial cells to make new blood vessels". Chronic skin wounds are an increasing medical problem, since they are commonly found in diabetic patients and in those suffering from morbid obesity. Diabetic patients may develop "diabetic foot", a condition whereby wounds do not heal leading, in the most severe cases, to amputation. Source: Instituto Gulbenkian de Ciencia
MICHAEL ROTHENSTEIN R.A. (1908-1993) Michael Rothenstein was the son of Sir William Rothenstein, the celebrated painter printmaker, writer and principal of the Royal College of Art. Michael was born in London and studied at Central School of Arts and Crafts. His studies were interrupted by a long illness (myxoedoema,a disease caused by underactivity of the thyroid gland). This caused a lengthy depression from which he eventually recovered. In 1940, he moved to Essex where he painted landscapes and rural subjects. In 1946 he embarked on a career as a printmaker and began to explore the many forms of print, including lithography, etching, aquatint, woodcuts, linocuts and screenprinting amongst others. He had numerous solo exhibitions throughout the world and his work is held by many public galleries such as the Tate ( London ), Museum of Modern Art ( New York ), British Museum, Smithsonian Institute (Washington) and the National Gallery of Australia . During a printmaking career that stretched over nearly half a century he earned a reputation as one of the most exciting and versatile printmakers of the century. He sometimes combined methods, such as woodcut, linocut and screenprint in a single print. Another of his contributions was incorporating found objects (such as bits of plank or crates, cogs, bits of metal) into relief prints and he published three books on his approach to printmaking. His subject matter was wide ranging but he often included recognizable, figurative elements and more abstract textural components in a collage-like composition. To find out more see: Tessa Sidey, (1993) The Prints of Michael Rothenstein, Scolar Press, Aldershot
# Question: Is x Even ## Comment on Is x Even ### For statement 2, I plugged in For statement 2, I plugged in numbers. And every time it was an even number for x and an odd number for y, the = was odd. Every time I flipped it, the = was even. So I concluded that x = even. Why would plugging in numbers not work here? ### Can you show me your Can you show me your calculations. I ask, because I think your calculations might be off. Statement 2: 6x - 3y is odd - case a: if x is even and y is odd, then 6x - 3y is ODD. In this case, x IS even - case b: if x is odd and y is even, then 6x - 3y is ODD. In this case, x is NOT even Both cases satisfy statement 2, but each case yields a different answer to the target question. This means statement 2 is not sufficient. ### can statement 2 be solved can statement 2 be solved this way: the difference btwn x and y means that x and y are not the same (one must be even, the other must be odd). Since we have no clue whether it's x or y that's even, since it works either way, statement 2 is not suff, so answer is A ### Be careful. Statement 2 does Be careful. Statement 2 does talk about the difference between x and y; it talks about the difference between 6x and 3y. Since 6x must be EVEN, we can conclude that y is odd. If the target question had asked "Is y odd?", your approach would have found statement 2 to be insufficient, when it would have been sufficient. ### For statement 1, if I look at For statement 1, if I look at xy+y is odd then either of the two terms will have to be odd. If we assume that y= odd, then x will have to be even since first term xy = has to even. Alternatively, if y=even, then x will have to be odd since first term has to be odd. This statement 1 comes out as insufficient. What am I missing here. ### Your first statement is true. Your first statement is true. However, your second statement is not true: Alternatively, if y=even, then x will have to be odd since first term has to be odd. If y is even, then the first term (xy) will be even, regardless of the value of x. This tells us that, if xy+y is odd, then y must be odd. Cheers, Brent
मुंबई में हर साल लगभग ५० से अधिक क्रूज आते हैं। यही कारण है कि क्रूज टूरिज्म की क्षमता को देखते हुए मुंबई से गोवा तक 'आंग्रिया' नामक क्रूज भी चलाया जा रहा है। कई सुविधाओं से परिपूर्ण अंतर्राष्ट्रीय क्रूज 'कोस्टा लुमिनोसा'(कोस्टा ल्युमिनोसा) सोमवार को मुंबई पहुंचा। ११० दिनों के वर्ल्ड टूर पर निकला यह क्रूज २६ मार्च को सिंगापुर से रवाना हुआ था। यह इस समय मुंबई के इंटरनेशनल क्रूज टर्मिनल पर डेरा डाले हुए हैं। आपको १४ मंजिला कोस्टा लुमिनोसा क्रूज में लगभग ३००० लोग यात्रा कर रहे हैं। यह क्रूज मुंबई होते हुए मालदीव जायेगा। बताया जाता है कि इस क्रूज में अंतर्राष्ट्रीय स्तर के सभी लक्जीरियस साजो सामान उपलब्ध हैं, इस क्रूज में कुल १,१३0 केबिन्स हैं। १20 मुरानो कांच के झूमर लगे हुए हैं। इसके अलावा ३ स्विमिंग पुल, ५ रेस्टोरेंट, डिस्को और सिनेमागृह भी उपलब्ध है। यह क्रूज किसी बहुमंजिला इमारत से कम नहीं दिखाई देता। इस क्रूज में क्रू मेंबर सहित लगभग ३ हजार लोग सवार हैं इसीलिए इसे 'छोटा गांव' कह कर संबोधित किया जा रहा है। सूत्रों के मुताबिक अब भारत में भी अंतर्राष्ट्रीय स्तर का क्रूज टर्मिनल विकसित किया जा रहा है ताकि विदेशो से आने वाले बड़े-बड़े क्रूज को भारत में रुकने के लिए कोई परेशानी न हो। मुंबई पोर्ट ट्रस्ट के अध्यक्ष संजय भाटिया ने बताया कि मुंबई में हर साल विदेशों से क्रूज आने की संख्या बढ़ रही है। इसीलिए मुंबई के क्रूज टर्मिनस को अत्याधुनिक बनाने का जो कार्य किया जा रहा है। यह कार्य इस साल के आखिर तक पूरा हो जायेगा। गौरतलब है कि मुंबई में हर साल लगभग ५० से अधिक क्रूज आते हैं। यही कारण है कि क्रूज टूरिज्म की क्षमता को देखते हुए मुंबई से गोवा तक 'आंग्रिया' नामक क्रूज भी चलाया जा रहा है।
#fm-pw6-forbiddenisle aka TomBot - Bot for the Freies Magazin 6th programming challenge # Copyright (C) 2013 Thomas Pummer # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. class rawInputReader(object): ## NOUNITTEST: there's only a raw_input in here def read(self): return raw_input() class fakeInputReader(object): def __init__(self, input): self.linecount = 0 self.lines = input def read(self): try: result = self.lines[self.linecount] except IndexError: result = 'IndexError' self.linecount = self.linecount + 1 return result def coordinateTextToTuple(inputCoordinates): coordinates = inputCoordinates.split(',') for i in range(len(coordinates)): coordinates[i] = int(coordinates[i]) - 1 return (coordinates[0], coordinates[1])
Peripheral Nerves Disorders The peripheral nervous system is the division of the nervous system that consists of the nerves and ganglia on the outside of the brain and spinal cord. The primary function of the peripheral nerves is to connect the central nervous system to the limbs and organs, mostly serving as a communication relay between the brain, spinal cord and the rest of the organism. Incoming and outgoing impulses travel in the nerves similar to a telephone wire. Sensory nerves conduct information about the environment such as temperature, touches, pain and the position of extremities. Motor nerves send bioelectrical impulses to muscle, resulting in contraction and movement. Unlike the brain, the peripheral nerves are not protected by the bone of spine and skull, or by the special tissues barrier between blood and brain. Therefore, it more exposed to toxins and mechanical injuries. Neuropathy is a condition that prevents nerves from working properly. It can cause paralysis in a case of complete laceration of nerve, although total paralysis is uncommon in patients with neuropathy. Rather, the disorder causes varying degrees of weakness and insensibility, depending on the neuropathy’s type and severity. Peripheral neuropathy includes damage to the peripheral nerves that transmit pain and temperature sensations, and can prevent sensing that patient has been injured from a cut or other causes. Besides, pain receptors in the skin can also become over-sensitized. Therefore people may feel severe pain from normal stimuli. There are numerous illnesses which involve peripheral nerves system. Some common examples of them include: • Guillain-Barre´ Strohl Syndrome (It is one of the most common causes of acute neuromuscular paralysis. A severe, rising, and progressive neuropathy characterized by weakness and paresthesias) • Chronic Inflammatory Demyelinating Polyneuropathy (CIPD) • Diabetic Neuropathies • Mononeuropathies (an impairment of a single nerve, typical example is Carpal Tunnel Syndrome – isolated numbness of the hands which can be caused by excessive keyboard work. • Peripheral Nerve Injuries • Amyotrophic Lateral Sclerosis (ALS): the Disorder of the motor nerves resulting in progressive weakness of muscles. It is the most dangerous and severe neuromuscular disorders. • Small Fiber Neuropathies • Occupational Neuropathies: Industrial and athletic injuries to nerves result in arm weakness and tingling. Peripheral neuropathy can be caused by: - Nerve compression or entrapment - Trauma and injuries - Fracture or dislocated bones - Tumor (including tumors of surrounding tissues (fat, muscle, vessels) which can compress the nerve - Intraneural hemorrhage - Prolonged exposure to cold or radiation - Certain medicines or toxic substances - Vascular or collagen disorders such as atherosclerosis, lupus, scleroderma, sarcoidosis, and rheumatoid arthritis. - In some cases, neuropathy is due to heredity, vitamin deficiency, infection, and kidney disease. The most common symptoms of peripheral neuropathy include weakness, muscle cramps or twitching, pain without obvious causes, numbness, burning and tingling. Symptoms depend on the type of affected nerve and may be observed over a period of days, weeks, or years. Neuropathic pain is hard to control with medicines and can severely impair emotional well-being and life quality. Neuropathic pain is frequently worse at night, seriously disordering sleep and adding to the emotional burden of sensory nerve injury. Motor nerve disorders cause muscle weakness, including painful cramps and twitching, muscle degeneration, and changes in the skin and hair. Sensory nerve disorders may result in a general feeling of numbness, mainly in the hands and feet. Damage to these nerve fibers may cause insensitive to injury from a cut or that a wound is becoming infected. Moreover, patients with peripheral neuropathy may not detect pains that signal of impending heart attack or other acute conditions. Also, pain receptors can become oversensitized and that patients may feel severe pain from routine stimuli. Neuropathy is difficult to diagnose, and it requires a thorough and comprehensive examination of all organ systems. Doctor will take a full medical history and perform a physical and neurologic examination that may include: - Checking of tendon reflexes - Checking of muscle strength and tone - Ability to feel certain sensations, posture and moves coordination Additionally, doctor may request the following tests: - Blood tests to check level of vitamin B12 - Thyroid function tests (blood tests to check level of thyroid hormones) - Electromyography (EMG) a procedure that measures the electrical discharges produced in muscles - A nerve conduction study, which estimates how quickly nerves transmit electrical signals. A nerve conduction study is usually used to diagnose carpal tunnel syndrome and other peripheral nerve diseases - In rare cases, a physician may prescribe a nerve biopsy, a procedure in which a small sample of a nerve is extracted and examined for malformations. However, even a nerve biopsy may not always explain the cause of some neuropathies. Neuropathy does not pass away unless the underlying disorders are relieved or removed. Although controlling a chronic condition may not reduce neuropathy, it can play an essential role in managing of the neuropathy. Depending on the cause, neuropathy may be treated by pharmacotherapy, vitamin or other nutrients supplements, physical or occupational therapy, surgery. The most important thing is treating of underlying conditions. For example: Diabetes. In the case of diabetes, a patient with his physician must “work together” to keep blood sugar level as close to normal ranges as possible. Maintaining normal blood sugar levels supports protection of nerves. Vitamin deficiency. If neuropathy is the outcome of a vitamin deficiency, a doctor may prescribe injections of the certain vitamin. In the case of pernicious anemia, a patient needs regular injections of B12-vitamin for the rest of life, and perhaps additional vitamin supplements. Autoimmune disorder. If neuropathy is caused by a prolonged inflammatory or autoimmune process, neuropathy treatment will be aimed at normalization and modulation of the immune response. Nerve pressure. Treatment will likely concentrate on adding ergonomic positions, desks or keyboards to home or office, changing the way of holding tools or instruments, or taking a break from certain activities or sports. In extreme cases of nerve compression, surgery is requered. Toxic substances or medications. If toxins or some drugs caused the neuropathy, it is critical to avoid further exposure to that toxins. Peripheral nerve dysfunctions caused by tumors or traumas as well as compressive conditions can sometimes be treated surgically. Regarding diabetic neuropathy, surgeons may consider implantation a spinal cord stimulator if a patient has medically refractory pain.
Why Do Buds Fall Off Lemon Tree Before Flowering I have a eucalyptus lemon tree and it's five months old. It's budding but the buds fall off before flowering. Is this normal, and if not, why does it happen? It's kept indoors at present. Certified GKH Gardening Expert If it is only 5 months old, the tree is probably purging the buds because it is still too small to support the fruit the buds would produce. This is normal. In fact, I would recommend that until the tree is 2 years old, that you remove any blossoms it produces so that it can focus on becoming well established rather than on growing fruit. It will be better for the long term health of the tree.
# What is half of 5 2? What is half of 5 2? The correct answer is -10.5 — notice the space between the 2 and the 5. And it’s quite informative to know that 62/5 is 10.5. (What is half of 5 2?) #### What is the half of 5 by 4? However, none of these can be the half of 5 by 4. This is because when solving a fraction, we multiply the numerator by the denominator and get the numerator and denominator of the solution together. In a fraction, the numerator and denominator cannot have the same value. (What is half of 5 2?) #### What is 2 divided by 5 as a fraction? To find the fraction of 2/5, we need to find the numerator and the denominator. 2/5 = x/5. Here x is the numerator. The denominator is calculated by (total number of numbers in the denominator) – (numerator). We have to find the number of numbers in the denominator. It is 10. (The number of numbers in the denominator) – (numerator) = 10-x. x = 10-x. x = 10-3. 3 is the denominator of 2/5. #### What’s half of 5 in a fraction? In a fraction, the denominator determines the number of equal parts that the whole or value of the fraction is divided into. So if the denominator is 5, then the number of equal parts is 10. If the denominator is 25 (or 1/4), then the number of equal parts is 5. #### What is half of 2 5 fraction? Although numerals are often verbalized as words, it is only to clarify the pronunciation of the words that follow. If a numeral is used to express a measurement, it is always written as numerals. For example, like 2 inches, 3 miles, 5 litres. If a fraction is used as part of a two-part business name, it is written as numerals. For Example, as 2H Supply Company, and 3H Printing Shop. If a fraction is used in a question, the answer is written as numerals. For example, if the question is: What is 3/2 of 1/8? The answer is 1/4. #### What is ¼ divided by 4? Ok, let me explain it to you this way. We know that 2 5 fraction is, as you said, 20/5. To get the one half of that, we will multiply the numerator and the denominator by 2. If the numerator is multiplied by 2, then we have the same number of zeroes. To make the denominator have the same number of zeroes, we have to divide it by 4. So the answer is that one-half of the 2 5 fractions is 10/5. #### What does 5 divided by 2 look like? This is a hard question.  If you’re a teacher, a parent, or a math whiz, you probably didn’t have much of a problem understanding this question. But chances are, you’re not. #### How do you divide fractions with whole numbers? You can divide fractions by multiplying them by the reciprocal of the divisor. You can also divide them by using long division. But dividing by multiplying is usually faster. Example: Divide 3/8 by 5/8. Multiply top by bottom: 3/8 *5/8 = 15/32 In long division, you find the divisor in the divisor column (5), and you divide the numerator by the same number in the dividend column (3). It’s called long division because you do it by columns instead of the traditional calculator method. #### What can you divide by 7? By subtracting 7 from 21, you get 12. You can divide 12 by 7 by dividing 12 by 7. This is a pretty easy question. You divide it the same way you would divide any number by 7. The answer is 3. In other words, if you had 12 apples, but you lost 7 of them, then you’d have 3 apples left. Or if you had 12 dollars, but you spent 7 dollars, then you’d have 3 dollars left. #### Can you divide 5? The short answer is no, you can’t. The language has some fundamental flaws in the way it is designed. For example, consider the fact that there is no symbol in the language to represent zero. So, how can you deal with division then? #### What is a number divided by 6? The correct answer is zero. Right? Well, not always. There can be exceptions to everything. If you had just read this question, and someone asked you the answer out of the blue, then your first reaction would be to respond zero. But did you know that there is a special case where you would still answer zero? #### How do you divide 2? First, you have to look at the order of numbers. If the 2 is in front of the 1 then the answer is 2, if the 2 is after the 1 then the answer is 1. In this case, the 2 is after the 1 so the answer is 1. If you have a question like 324 divided by 2 you have to look at the last number and start dividing from there. Start dividing by 2’s until you get to the number one, then carry the one over to the other column. This will give you 2 as the last number which means the answer is 2. #### What is the third of 6? The third of six is the first part of a series of three. It is the third of the second part. The third of six is an example of a series of three. It is an example of the first part of a series of three. The third of six is a representation of three. The third of six is a part of the second part of a series of three. #### What is the 48 split into 6? A 48-split is an N-digit game where every combination for N digits is assigned an N-digit number. Now, in your question, N=6, so N digits will be 6 digits in total. #### Is there a division table? When you’re planning your social media strategy, keep in mind that Facebook isn’t the only game in town — and for some businesses, it’s not even a game. According to a comprehensive survey of small businesses and social media, only % of businesses use Facebook (compared to % using Twitter and __% using LinkedIn). So while Facebook should most definitely be a part of your overall social media strategy, you should also explore other platforms. #### What can 24 be divided by? This question is a classic example of an impossible problem. There is no solution because there is no whole number that evenly divides into 24. There are multiples that come close. The closest is 12, which is 62, and the next closest is 16, which is 82. However, neither one is an exact multiple of 24. Other close multiples are 17, 20, 25, 28 and 30. #### What is the answer for 2 divided by 2? 2 divided by 2 can be written mathematically as follows: 2 ÷ 2 = 1. The answer 1 is an integer, meaning that it is an accurate representation of 1 object divided into exactly 2 pieces. The problem of dividing 1 object into 2 pieces can be thought of as an equivalent problem: putting 2 objects into 1 bag. #### How do u divide fractions? You CANNOT divide fractions. Fractions represent an amount left over after a division problem. So let’s say you have a whole piece of the pie. Well, what if 2 people want a piece of the pie. You can cut the pie into 4 pieces and give the 2 people 1/4 each. Now, what if the people wanted the same piece? You would give each of them 1/2 of the pie which would leave you with 1/4 of the whole pie. Since 1/4 is left over, it would be represented as a fraction: 4/4 = 1/4. Fractions are used to show how much of a whole is left over, not how much of a whole is covered.\
A new paper from an assistant professor at the University of California, Irvine details the potential effects that a solar superstorm could have on our heavily internet-reliant world. The conclusion? The effects of a particularly strong solar storm could have devastating effects on undersea internet cables, a crucial component of the world's internet infrastructure. Without stronger mitigation efforts against these effects, the study claims we could be headed towards an "internet apocalypse." To read more, click here.
आप साहसी प्रवृत्ति के व्यक्ति हैं किसी भी कार्य को उत्सुकता पूर्वक करने वाले होते हैं तथा आपके हर कार्य में जिम्मेदारियां झलकती हैं। नेतृत्व की चाह आपको अच्छी उन्नति दिला सकता है। आप अपने कार्य शैली की वजह से जाने जाते हैं। समय व परिस्थिति के अनुसार कार्य करने वाले होते हैं। परंतु जल्दबाजी में कार्य करने से या कोई ठोस निर्णय लेने से नुकसान होने की संभावना ज्यादा रहती हैं। यदि आप किसी भी कार्य को सोची समझी रणनीति के तहत करने का प्रयास करते हैं तो आपको अच्छी कामयाबी देखने को मिल सकती हैं। क्योंकि आपके अंदर किसी भी कार्य को सफल बनाने की साहस है। आप अपने आत्मविश्वास को जागृत करते हुए अपने मंजिल को प्राप्त कर सकते हैं। इस माह में आपको धन रिलेशन तथा किसी नए व्यक्ति से संबंध जुड़ने पर कुछ नया कार्य व्यवसाय में इजाफा हो सकता है। इस माह में किए गए कार्यों से अच्छी सफलता प्राप्त होने की संभावना बन रही हैं। यदि आप अपने भाइयों के साथ मिलकर कोई कार्य करते हैं तो आपको कामयाबी अच्छी मिल सकती है। इस माह में उन्नति के लिए अच्छे अवसर प्राप्त हो सकते हैं। चाहे आप नौकरी करते हो या व्यवसाय करते हो फिर भी आपको अच्छा नेतृत्व करने का मौका मिल सकता है। पद पोजीशन प्राप्त होने की संभावना अच्छी पाई जाती है। यदि आप राजनीतिज्ञ है तो इस माह में सावधानी बरतते हुए किसी भी कार्य को करने का कोशिश करें। इस माह में जनसमर्थन प्राप्त होने की संभावना पाई जाती है। आपको अपने विरोधियों का शिकार होना पड़ सकता है। यदि इस माह में कोई गाड़ी या घर लेने का प्रयत्न है तो अभी इंतजार करना चाहिए। माता-पिता के स्वास्थ्य को लेकर चिंता उत्पन्न हो सकती हैं। अतः माता-पिता के स्वास्थ्य के प्रति जागरुक रहें। किसी तरह का कोई घरेलू प्रॉपर्टी को लेकर या किसी अन्य तरह का कोई विवाद चल रहा हो तो इस माह में समझाने का प्रयास करना आपके लिए अच्छा हो सकता है। पारिवारिक उलझने कम रहने की संभावना पाई जाती हैं। किसी तरह की कोई शत्रुता उत्पन्न न हो इसका भी ख्याल रखना जरूरी है। स्वास्थ्य को लेकर कोई गंभीर समस्या नहीं है। बाहर की यात्रा तथा दांपत्य जीवन बेहतर स्थिति में हो सकता है। वैवाहिक जीवन सफल होने के साथ साथ बाहर की यात्रा भी सफल हो सकती है। इस माह में भाग्य आपका साथ कम देगा। आपका कर्म ही आपका प्रयत्न ही आपको कामयाबी दिला सकती है। भाग्य के सहारे रहना या कोई कार्य करना आपके लिए अच्छा नहीं है। आप संघर्षशील व्यक्ति हो इसलिए संघर्ष करने से आपको अच्छी कामयाबी मिल सकती है। क्योंकि करियर के दृष्टि से स्थिति काफी अच्छी है। आर्थिक लाभ के दृष्टि से भी यह माह उन्नति दायक हो सकता है। यदि किसी भी कार्य को बुद्धिमानी से किया जाता है तो वह सफल होने की संभावना बन सकती हैं। इस माह में २,३,११,1२, २0, २1 तारीख आपके लिए अनुकूल नहीं है। इस माह में मानसिक तनाव बढ़ने की संभावना है। अतः इस तारीख में किसी तरह का कोई ठोस निर्णय लेना किसी तरह के कार्य का शुभारंभ करना उचित नहीं रहेगा। सुलझे हुए व्यक्ति के तरह कार्य करने से व्यक्तित्व में भी निखार आता है तथा कार्य व्यवसाय में भी तरक्की होती है। आप भी ऐसा कर सकते हैं। आर्थिक दृष्टि से इस माह में कुछ विस्तार योजनाओं को मंजूरी मिल सकती है। उससे अच्छा आर्थिक लाभ हो सकता है। कामकाज को लेकर स्थितियां काफी अच्छी रहने वाली हैं। इसलिए आप उम्मीद कर सकते हैं कि इस माह में आर्थिक लाभ के लिए किया गया प्रयास आपके लिए हर तरह से उपयोगी हो सकता है। भाग्य आपका बहुत अच्छा साथ नहीं देगा। इसलिए आपको सोच समझ कर किसी भी तरह का आर्थिक इन्वेस्ट करना चाहिए जिससे कि आपको कोई नुकसान न हो। इस माह में धन अचल संपत्ति के योग अच्छा बन रहा है। अतः किसी तरह का कोई धन संपदा को लेकर इन्वेस्टमेंट करना पड़े तो आप कर सकते हैं। आपके द्वारा आर्थिक लेनदेन की प्रक्रिया को सुधार लाना तथा उसके प्रति जागरुक रहना आपके लिए लाभदायक हो सकता है। यदि आप किसी को धन देने का मन बना रहे हैं तो आपको विचार करना चाहिए। समय और परिस्थिति को देखते हुए धन का आदान प्रदान करना चाहिए। आर्थिक रुप से इस माह में स्थितियां मजबूत रहने वाली हैं। इसलिए अर्थ लाभ का प्रयत्न जारी रखना चाहिए और हर संभव आर्थिक लाभ के लिए प्रयास करना आपके लिए बेहतर होगा। जिस किसी कार्य को करने का मन बना रहे हैं उसे इस माह के उत्तरार्ध में कर सकते हैं। जिससे अर्थ लाभ अच्छा हो सके। किसी तरह का कोई गंभीर समस्या उत्पन्न होने की संभावना नहीं है। फिर भी यदा-कदा बुखार जुखाम इत्यादि हो सकता है। किसी तरह का कोई चोट चपेट इत्यादि की संभावना बन सकती हैं। अतः अपने स्वास्थ्य से संबंधित सावधानी रखना अनिवार्य है। इस माह में प्रेम संबंधों को लेकर स्थितियां पक्ष में हो सकती हैं परंतु किसी तरह के अनावश्यक बात विचार को लेकर तनाव उत्पन्न हो सकता है। किसी तरह के घूमने फिरने के उद्देश्य को लेकर तनाव उत्पन्न हो सकता है। ऐसे में एक दूसरे के प्रति अच्छा सामंजस्य बना कर बातों से उसका हल निकालना ज्यादा बेहतर हो सकता है। इस माह में आपको अपने प्रेमी/प्रेमिका को प्रसन्न रखने का प्रयास करना ही आपके लिए बहुत बड़ा योगदान हो सकता है। यदि इस माह में आप अपने प्रेमी/प्रेमिका को खुश रखने की कोशिश करते हैं तो आपके कार्य व्यवसाय में भी अच्छा सहयोग प्राप्त हो सकता है। आप हमेशा खुश रहकर अपने कार्य व्यवसाय में भी अच्छी इजाफा कर सकते हैं। हो सके तो इस माह में अपने प्रेमी/प्रेमिका से कुछ गुप्त चर्चाएं न करें। ऐसा भी हो सकता है कि आप जिस किसी को प्यार करते हैं वह आपके वैवाहिक जीवन में परिवर्तित होने के लिए भी इच्छा जाहिर कर सकता है। इस माह में आपका वैवाहिक जीवन सफल होने की संभावना पाई जाती है। पति/पत्नी के अटूट प्रेम से हर किसी मुसीबतों से मुकाबला किया जा सकता है तथा हर तरह के तनाव को दूर किया जा सकता है। बाहर की यात्रा तथा कामकाज के क्षेत्रों में जीवनसाथी का सहयोग प्राप्त होने की संभावना बन सकती हैं। ऐसा भी हो सकता है कि यदि आपको किसी तरह का कोई सहयोग चाहिए तो वह ससुराल पक्ष से प्राप्त हो सकता है। इसलिए जीवनसाथी के साथ मधुर संबंध बनाए रखना आवश्यक है। इस माह में पारिवारिक स्थितियों को लेकर तनाव में हो सकते हैं। घर परिवार में हर तरह की समस्याएं देखने को मिल सकती हैं। जिन समस्याओं को निपटाने का प्रयास करेंगे वह उतना ही जटिल होता जाएगा। किसी तरह के लेन देन के मामले को लेकर या किसी तरह के छोटी-छोटी समस्याओं को लेकर बड़ी समस्या उत्पन्न की जा सकती है। किसी अन्य व्यक्ति की दखल से भी समस्या उत्पन्न हो सकती है। घर परिवार में आपसी सामंजस्य का अभाव होने के कारण इस तरह की मुसीबतें ज्यादा हो सकती हैं। इसलिए घर परिवार में सबके साथ आपसी सामंजस्य बेहतर हो इसका भी ख्याल रखना आवश्यक है। अन्यथा घरेलू विवाद के चलते पारिवारिक कलह बढ़ सकता है। कार्य व्यवसाय पर भी इसका बुरा असर पड़ सकता है। अतः इन चीजों का ख्याल रखें। माता- पिता के स्वास्थ्य के प्रति चिंता करना तथा उनके स्वास्थ्य को लेकर जागरुक रहना भी आपके लिए जरूरी है। हो सकता किसी तरह की कोई समस्या उत्पन्न हो। भाई बहन छोटे बड़े सबको एक साथ एकत्र करने की कोशिश करें। संतान पक्ष को लेकर स्थितियां अनुकूल रहने की संभावना बन रहे हैं। संतान की पढ़ाई लिखाई को लेकर बेहतर महसूस कर सकते हैं। घर परिवार की भी जिम्मेदारी बहुत बड़ी जिम्मेदारी होती है और इस जिम्मेदारी को भी निभाना बहुत जरूरी होता है तभी इंसान आगे अच्छी उन्नति कर सकता है। अतः इसका भी ख्याल रखें। आप सूर्य को निरंतर जल दें तथा आदित्य हृदय स्तोत्र का पाठ करें। रविवार के दिन लाल वस्तुओं का दान करें तथा गरीब व्यक्तियों की सेवा करें। ऐसा करने से आपकी शारीरिक समस्याएं कम हो सकती हैं। काम काज के क्षेत्र में भी लाभ प्राप्त हो सकता है।
The appropriate metrics to manage and measure the success of a company’s operation vary significantly by industry, by individual company, and by the scale of the business. What does not vary, however, is the universal need of all companies to employ a well-structured, hierarchical framework to organize and manage their metrics. The absence of a cohesive framework to house metrics greatly increases the likelihood that a company’s performance measurement system (PMS) will provide inadequate management support, and that resources will be wasted developing duplicative, unaligned and even conflicting metrics. There are a number of well-known models and frameworks for operations, logistics and supply chain management. Two of the most prominent are the SCOR model and the Balanced Scorecard, and the interested reader is referred to these. Figure 1 depicts an integrated hierarchical supply chain performance measurement system. The framework contains three levels (the strategic, tactical and operational), and within each level, it has both external and internal measures. In this PMS framework, it is the scale of an operation or activity that a particular metric monitors which determines its place in the hierarchy. As we observe in Figure 1, metrics that measure the performance of an entire major functional area (such as distribution) are considered strategic, while metrics that monitor major sub-functions (such as warehouse operations and transportation) are categorized as tactical, and finally metrics that monitor sub-functions of sub-functions (e.g., the receiving operation in a warehouse) are considered operational metrics. One can adapt this PMS framework to as many or as few echelons as may exist in your company’s supply chain. Figure 2 provides additional insight on how this hierarchical PMS framework works, displaying sample external and internal metrics for a distribution organization at each level of the hierarchy. The external metrics measure outputs and/or services that flow across the supply chain and evaluate some aspect of serving the customer. The internal metrics have an “inward” focus; and as shown in Figure 2, they evaluate how efficiently the overall distribution organization and each of its sub-functions operates. Employing a hierarchical performance measurement framework offers a number of important benefits. First of all, this approach provides a unified framework for aggregating performance metrics across a company. By implementing a hierarchical measurement approach, a company positions itself to organize its key performance measurements into a structure that leads to a relatively few, high-level, strategic measures (e.g., between 10 and 25) that monitor overall corporate performance. At the same time, this structure facilitates having many other performance measures that monitor smaller components of a company’s operation, yet which align with overall objectives. In particular, a hierarchical measurement system allows both large and small functional areas within a company to: - develop and maintain their own measures, and - contribute to and be part of an overall measurement system. This hierarchical approach thus helps to keep measures both simple and meaningful because each function and sub-function at each level can focus on a few key performance measures. Finally, a hierarchical performance measurement system can also contribute towards aligning the collective activities of a company to meet a desired mission and set of objectives. For example, if a company has a comprehensive measurement system in place that covers most or all of its major functional areas and activities, managers can view the system in its entirety to identify any potential misaligned activities or objectives. Tan Miller is director of the Global Supply Chain Management Program at Rider University, College of Business Administration, and is a member of MH&L’s Editorial Advisory Board. Previously he worked in private industry for over 20 years where most recently he was responsible for the operations of J&J’s U.S. Consumer Distribution Network.
Metamath Proof Explorer < Previous   Next > Nearby theorems Mirrors  >  Home  >  MPE Home  >  Th. List  >  dmxp Structured version   Visualization version   GIF version Theorem dmxp 5376 Description: The domain of a Cartesian product. Part of Theorem 3.13(x) of [Monk1] p. 37. (Contributed by NM, 28-Jul-1995.) (Proof shortened by Andrew Salmon, 27-Aug-2011.) Assertion Ref Expression dmxp (𝐵 ≠ ∅ → dom (𝐴 × 𝐵) = 𝐴) Proof of Theorem dmxp Dummy variables 𝑥 𝑦 are mutually distinct and distinct from all other variables. StepHypRef Expression 1 df-xp 5149 . . 3 (𝐴 × 𝐵) = {⟨𝑦, 𝑥⟩ ∣ (𝑦𝐴𝑥𝐵)} 21dmeqi 5357 . 2 dom (𝐴 × 𝐵) = dom {⟨𝑦, 𝑥⟩ ∣ (𝑦𝐴𝑥𝐵)} 3 n0 3964 . . . . 5 (𝐵 ≠ ∅ ↔ ∃𝑥 𝑥𝐵) 43biimpi 206 . . . 4 (𝐵 ≠ ∅ → ∃𝑥 𝑥𝐵) 54ralrimivw 2996 . . 3 (𝐵 ≠ ∅ → ∀𝑦𝐴𝑥 𝑥𝐵) 6 dmopab3 5369 . . 3 (∀𝑦𝐴𝑥 𝑥𝐵 ↔ dom {⟨𝑦, 𝑥⟩ ∣ (𝑦𝐴𝑥𝐵)} = 𝐴) 75, 6sylib 208 . 2 (𝐵 ≠ ∅ → dom {⟨𝑦, 𝑥⟩ ∣ (𝑦𝐴𝑥𝐵)} = 𝐴) 82, 7syl5eq 2697 1 (𝐵 ≠ ∅ → dom (𝐴 × 𝐵) = 𝐴) Colors of variables: wff setvar class Syntax hints:   → wi 4   ∧ wa 383   = wceq 1523  ∃wex 1744   ∈ wcel 2030   ≠ wne 2823  ∀wral 2941  ∅c0 3948  {copab 4745   × cxp 5141  dom cdm 5143 This theorem was proved from axioms:  ax-mp 5  ax-1 6  ax-2 7  ax-3 8  ax-gen 1762  ax-4 1777  ax-5 1879  ax-6 1945  ax-7 1981  ax-9 2039  ax-10 2059  ax-11 2074  ax-12 2087  ax-13 2282  ax-ext 2631  ax-sep 4814  ax-nul 4822  ax-pr 4936 This theorem depends on definitions:  df-bi 197  df-or 384  df-an 385  df-3an 1056  df-tru 1526  df-ex 1745  df-nf 1750  df-sb 1938  df-eu 2502  df-mo 2503  df-clab 2638  df-cleq 2644  df-clel 2647  df-nfc 2782  df-ne 2824  df-ral 2946  df-rab 2950  df-v 3233  df-dif 3610  df-un 3612  df-in 3614  df-ss 3621  df-nul 3949  df-if 4120  df-sn 4211  df-pr 4213  df-op 4217  df-br 4686  df-opab 4746  df-xp 5149  df-dm 5153 This theorem is referenced by:  dmxpid  5377  rnxp  5599  dmxpss  5600  ssxpb  5603  relrelss  5697  unixp  5706  xpexr2  7149  xpexcnv  7150  frxp  7332  mpt2curryd  7440  fodomr  8152  nqerf  9790  dmtrclfv  13803  pwsbas  16194  pwsle  16199  imasaddfnlem  16235  imasvscafn  16244  efgrcl  18174  frlmip  20165  txindislem  21484  metustexhalf  22408  rrxip  23224  dveq0  23808  dv11cn  23809  mbfmcst  30449  eulerpartlemt  30561  0rrv  30641  noxp1o  31941  noextendseq  31945  bdayfo  31953  noetalem3  31990  noetalem4  31991  curf  33517  curunc  33521  ismgmOLD  33779  diophrw  37639 Copyright terms: Public domain W3C validator
#!/usr/bin/env python """ Entry point for bin/* scripts """ __authors__ = "James Bergstra" __license__ = "3-clause BSD License" __contact__ = "James Bergstra <[email protected]>" import cPickle import logging import os import base import utils logger = logging.getLogger(__name__) from .base import SerialExperiment import sys import logging logger = logging.getLogger(__name__) def main_search(): from optparse import OptionParser parser = OptionParser( usage="%prog [options] [<bandit> <bandit_algo>]") parser.add_option('--load', default='', dest="load", metavar='FILE', help="unpickle experiment from here on startup") parser.add_option('--save', default='experiment.pkl', dest="save", metavar='FILE', help="pickle experiment to here on exit") parser.add_option("--steps", dest='steps', default='100', metavar='N', help="exit after queuing this many jobs (default: 100)") parser.add_option("--workdir", dest="workdir", default=os.path.expanduser('~/.hyperopt.workdir'), help="create workdirs here", metavar="DIR") parser.add_option("--bandit-argfile", dest="bandit_argfile", default=None, help="path to file containing arguments bandit constructor \ file format: pickle of dictionary containing two keys,\ {'args' : tuple of positional arguments, \ 'kwargs' : dictionary of keyword arguments}") parser.add_option("--bandit-algo-argfile", dest="bandit_algo_argfile", default=None, help="path to file containing arguments for bandit_algo " "constructor. File format is pickled dictionary containing " "two keys: 'args', a tuple of positional arguments, and " "'kwargs', a dictionary of keyword arguments. " "NOTE: bandit is pre-pended as first element of arg tuple.") (options, args) = parser.parse_args() try: bandit_json, bandit_algo_json = args except: parser.print_help() return -1 try: if not options.load: raise IOError() handle = open(options.load, 'rb') self = cPickle.load(handle) handle.close() except IOError: bandit = utils.get_obj(bandit_json, argfile=options.bandit_argfile) bandit_algo = utils.get_obj(bandit_algo_json, argfile=options.bandit_algo_argfile, args=(bandit,)) self = SerialExperiment(bandit_algo) try: self.run(int(options.steps)) finally: if options.save: cPickle.dump(self, open(options.save, 'wb')) def main(cmd, fn_pos = 1): """ Entry point for bin/* scripts XXX """ logging.basicConfig( stream=sys.stderr, level=logging.INFO) try: runner = dict( search='main_search', dryrun='main_dryrun', plot_history='main_plot_history', )[cmd] except KeyError: logger.error("Command not recognized: %s" % cmd) # XXX: Usage message sys.exit(1) try: argv1 = sys.argv[fn_pos] except IndexError: logger.error('Module name required (XXX: print Usage)') return 1 fn = datasets.main.load_tokens(sys.argv[fn_pos].split('.') + [runner]) sys.exit(fn(sys.argv[fn_pos+1:])) if __name__ == '__main__': cmd = sys.argv[1] sys.exit(main(cmd, 2))
How Kyphoplasty Can Correct Spinal Compression May 25, 2017 What Is Kyphoplasty? Kyphoplasty is a minimally invasive medical procedure to treat osteoporosis. It involves inserting a balloon in a fractured spinal vertebrae and filling the space with bone cement. The procedure usually targets the thoracic region of the spine in the T1 through T12 vertebrae. It can also be used in the lumbar region in L1 through L5. Kyphoplasty reduces pain, stabilizes the vertebrae and restores the vertebrae to their normal height. Why Osteoporosis Causes Spinal Compression Osteoporosis is a disease that involves bones becoming fragile. As people age, their bones lose calcium and other minerals. The vertebrae then fracture. The spine shortens, causing a hump to form in the back. The condition is known as kyphosis. Fractures caused by weakening of the vertebrae are called compression fractures. They can take place suddenly. Sometimes the fractures cause no initial problems. Later, the fractures can cause pain slowly that gets worse with walking. Over time, the fractures often lead to severe back pain. The pain can strike in the middle or lower spine, on the sides of the body or in front of the spine. The pain caused by compression fractures is acute. It can take weeks or months to ease. The fractures can cause a person to lose as much as six inches over time. The pain can result in tingling, numbness, difficulty walking, and loss of function in the bowels and bladder. How Kyphoplasty Fixes Spinal Compression The agent to fix the problem is bone cement, also known as polymethyl methacrylate (PMMA). PMMA hardens soon after it is placed in the body, yet it does not stick to the bones. It has no adhesive properties. Instead, it acts as a space-filler. It causes the irregular bone surfaces that remain in the vertebrae to lock close together. A kyphoplasty procedure involves an anesthesiologist and a doctor. The doctor inserts the needle into the bone and inflates the balloon. The doctor injects bone cement into the balloon, using X-rays to check location, then withdraws the needle. Kyphoplasty usually takes less than an hour for one vertebrae. It can take longer to treat more than one vertebrae. After a kyphoplasty procedure, a patient is usually walking normally. He or she is in much less pain than before. If you are interested in receiving this quality pain treatment, talk to one of our experienced physicians about the risks and benefits of kyphoplasty. The doctors at Southside Pain Specialists are happy to help you learn how to restore your height, health and ability to walk.
एक औरत की डायरी.........: जीने की तमन्ना संगीता स्वरुप ( गीत ) ७ सिप्तंबर २०१२ एट १४:२९ मन के द्वंद्व को बखूबी लिखा है बहुत सुन्दर सोनिया.... सुन्दर कविता ,काबिले तारीफ । बुरांस के फूल तो बहुत सुन्दर लगाए हैं. कविता भी सुन्दर है. बहुत खूब ... दिल को छू के गुज़र जाती है रचना ... जीने की तमन्ना तुम में नजर आती है...:) बहुत प्यारी... दिल से कही हुई रचना.. सोनिया जी, बहुत खूबसूरत रचना. आप सभी का बहुत बहुत शुक्रिया.... आप सभी के कमेंट मुझे आगे बढ्ने के लिए प्रेरित करते हैं अपना स्नेह सदा यूं ही बनाए रखे। ताकि मेरी लेखनी यूं ही चलती रहे।
The basic aim of this article is to highlight the role of legal transfer in European Post-Napoleonic constitutionalism. The “model-effect” of the Charte constitutionnelle as a prototype of “monarchical constitutionalism” will be analysed by focusing on one practical example, namely the Bavarian Constitution of 1818, which is generally seen as a “derivative” of the 1814 French Constitution. Such an understanding, however, does tend to be biased. Based on results from both comparative history and transfer research, the actual relevance of the French and other contemporary constitutions for Bavaria is evaluated, thus providing an analytical basis to gain further insight into the nature, scope and limits of constitutional transfer and reception at a more general level. In conclusion, existing concepts of constitutional “models” are questioned and the need to reassess the approach to transfer history is stressed both with regard to its methodology and content.
LAWRENCE — When growing up, many young people look to their parents for guidance on how to prepare for all manner of experiences. But when the parents do not have a particular experience to share, where do kids turn? A new study from the University of Kansas has found that, for students whose parents did not attend college, they turned to media of all sorts to form their expectations of college. And the expectations did not always match reality. Some research has shown that as many as one-third of students today are first-generation college attendees, yet they only graduate at about half the rate of their peers whose parents have a college degree. Joseph Erba, assistant professor of journalism & mass communications; Angela Gist-Mackey, assistant professor of communication studies, and Marissa Wiley, doctoral candidate in communication studies, authored a study in which they conducted in-depth interviews with 28 first-generation college students about their media consumption, expectations for college and experiences in their first year. “Compared to their peers, first-generation college students cannot ask their parents about what college life is like. Most of our participants’ parents never went to college, and none graduated,” Erba said. “We wanted to find out where these students turn to get their information on college life and what role media played in setting their expectations.” The findings showed that participants based assumptions on media of all types, including movies, television, social media, college brochures and schools’ websites. The majority of the respondents reported those assumptions did not match their experience three months into their first year, when the interviews were conducted. They also emphasized that they relied more heavily on media they perceived as credible, such as official university websites, brochures and social media channels. “We learned that media was critical to the way students sought information about college life,” Gist-Mackey said. “They used Google searches, YouTube and other types of social media to seek information about and form impressions about both academic and social experiences in college. Unfortunately, much of the media they consumed represented college as a huge party scene, which they found out was unrealistic to their actual experience in college.” While students were both disappointed and relieved that college is not, in fact, one giant party, nearly all reported picking up that expectation. Those who were disappointed gave examples such as how their dorm rooms lacked a doorstop to keep their rooms open to friends and passers-by. Those who dreaded a constant party atmosphere said they were glad they weren’t forced to socialize when they didn’t want to and were able to find time alone. Both groups were able to differentiate what they saw in movies as likely not realistic, but they still had the expectation of a near-constant party scene because they picked it up in several other forms of media. First-generation students’ expectations of academic life also did not necessarily match their assumptions. A majority of students reported being surprised that their classes were more difficult than they expected and that none of media they consumed set an expectation for the amount of work and study that would be required. Regardless of the type of media consumed, it played a much larger role in setting expectations than any other source of information, the researchers found. Students turned to all manner of media before teachers, school counselors or even older peers or former classmates who reached college before them. When they did reach out, one very specific form of media played a key role. “The students followed colleges through social media channels to get an idea of what life there might be like,” Erba said. “Some tried to connect with current students through social media as well if they had friends at the school they were interested in. Overall, media representations did influence their preparation and expectations. Many students were a little shocked by the differences in what they found.” The students reported that even representations they expected to be credible, such as social media posts by current students and representations in schools’ own materials and channels, did not always match reality. College brochures are noted for emphasizing, or in some cases overemphasizing, racial diversity on campus. Some study participants reported expecting a high level of diversity on campus, yet they found that wasn’t necessarily the case. Others reported expecting to experience high levels of school spirit on campus and ample opportunities to take part in athletics events, yet they didn’t necessarily find that upon arrival. Having a deeper understanding of how students use various media platforms to build college expectations could be beneficial in higher education. “The findings of this research will prove useful for many years to come, particularly as we continue to rely on technology and immerse ourselves deeper into the digital world. We found that students are seeking messages from multiple platforms, searches and connections online,” Wiley said. “Students are constantly bombarded by messages from various online platforms, so it is imperative that high schools and universities seek ways to improve messages produced for an authentic online presence across a multitude of online platforms. While first-generation college students were the main focus of our current research, universities should consider creating more realistic media campaigns that would help better prepare and situate expectations of collegiate life for all incoming students.” The findings also can provide insight into how to better reach and prepare such students for life in college, which can lead to better success. “Knowing how first-generation college students are seeking information about colleges and universities and the reality that most of them are feeling misinformed about college life can help universities create more authentic messages and determine what media touchpoints to use when attempting to access this population,” Gist-Mackey said. “High school and college practitioners need to collaborate more directly to reach these students and their parents before, during and after the college application and admissions processes. Institutions of higher education need to work to construct more authentic and meaningful messages about college life so that first-generation students can have better-informed impressions of what college life will be like.” The authors will present the study at the National Communication Association conference this month. “We want students to be prepared for college life,” Erba said. “The better prepared they are, the better they do. This could help lead to enhanced graduation rates and better experiences for students. It could be a win, win, win.”
|Edmunds, Peter J.||California State University Northridge (CSU-Northridge)||Principal Investigator| |Lasker, Howard||State University of New York at Buffalo (SUNY Buffalo)||Co-Principal Investigator| |York, Amber D.||Woods Hole Oceanographic Institution (WHOI BCO-DMO)||BCO-DMO Data Manager| Abundance of sponges (on both density scale – number per quadrat – and volume scale – estimated volume per quadrat) for three sites over time on the shallow coral reefs of St. John from 1992-2017. Data support Figure S2 in the paper published in Marine Biology (Edumunds et al., 2020). Other datasets from Edmunds et al. (2020) are listed on the page https://www.bco-dmo.org/related-resource/819411. The following methodology was extracted from Edumunds et al. (2020), figure and table references below correspond to that publication. This methodology section applies to this dataset in addition to other datasets published in Edmunds et al. (2020). Sponge abundance was determined using photoquadrats recorded annually since 1992, but here they were sub-sampled by years scattered uniformly across the 26 year study (described below). These photoquadrats are part of a time-series that was designed to quantify the coral community (Edmunds 2013, 2019). The project consists of six sites on hard substrata at 7–9-m depth that were randomly selected in 1992, permanently marked, and recorded using photoquadrats (0.5 × 0.5 m) positioned randomly along a single transect at each site. Photoquadrats were recorded in May 1992, June 1993, August 1994, May 1995-1997, and July or August thereafter. Prior to 2000, ~ 18 photoquadrats site-1 were recorded annually using 35-mm film (Kodachrome 64), but from 2000, ~ 40 photoquadrats site-1 have been recorded using digital cameras ranging in resolution from 3.3 MP to 36.3 MP. Cameras were fitted with two strobes (Nikonos, SB 105), and color slides were scanned (4000 dpi) for analysis. Definitive identification of Caribbean sponges requires analysis of spicules and other skeletal characteristics (Rützler 1978). As this was impossible with photoquadrats, a consensus set of 23 sponges was identified using expert opinion. This set resolved 19 species, 3 sets of congeners that could not be distinguished to species, and an unknown category (Table S1). Most sponges were identified to species (66%, n = 9,608 pooled over the whole study), 4% were identified to genus, and 31% were assigned to the “unknown” category. Surveys on the adjacent island of St. Thomas suggested the total sponge diversity includes over 100 species (Gochfeld et al. 2020). Between 100 and 300 sponge species is typical for a localized area in which multiple sites are combined (Wulff 2016), with individual Caribbean sites being characterized by 51–67 species (Villamizar et al. 2013, Wulff 2006b, 2013). To evaluate sponge density, photoquadrats were opened in Adobe Photoshop CS5.1 software, and the number of sponges counted. Individuals were defined by their contiguous areas of sponge biomass, which were separated from other sponges of the same taxon. This method potentially overestimates sponge density where algae, other taxa, or topographically complex surfaces obscured connections among pieces of a contiguous sponge. The limitations of planar photographs in quantifying benthic communities have been inherent in the method since it was first used in modern ecology (Loya 1972). These issues are unavoidable with planar images, but their magnitude depends on the quantity of macroalgae and the rugosity of the benthic surfaces. With regard to rugosity, the study sites were located on carbonate pavement (RS11, 5), or igneous substrata (RS15, 2), or a combination of the two (RS9, 6) (Fig. 1), which provided relatively smooth surfaces with limited ability to obscure sponges (Fig. 2). Analyses of rugosity at Europa Bay (RS11) and East Cabritte (~ 500 m from RS2) in 2014 revealed that the mean (± SE) topographic complexity at ~ 8-m depth was 1.16 ± 0.02 and 1.22 ± 0.02, respectively (Tsounis et al. 2018). The values for topographic complexity from St. John describe relatively flat communities (Fig. 2) that correspond to 40 year minimal values across the Caribbean, where rugosity declined from ~ 2.5 in 1969 to 1.2 in 2008 (over a depth range from the surface to > 20 m) (Alvarez-Filip et al. 2009). Evaluating the ecological importance of sponge density (i.e., sponges per area) requires measurements of sponge biomass (Wulff 2001, 2016), which are unobtainable from planar photographs. However, sponge biomass can be estimated from sponge volume calculated from linear dimensions and the volumetric formulae of geometric shapes matching those of the sponges (after Wulff 2001). We sought approximations of sponge volumes using photoquadrats, assuming that the shape of each sponge could be inferred from a planar image and volumetrically approximated by geometric shapes (i.e., rods, disks, and spheres). The volumes of these shapes were estimated using linear dimensions obtained from the photographs using ImageJ software (Abramoff et al. 2004). In the few cases where sponges were conical (e.g., Ircinia campana), it was not possible to accurately estimate their volume and they were excluded from the analysis. As I. campana accounted for only 1% of all sponges, the bias attributed to this affect was trivial. The volume of encrusting sponges was estimated assuming they were disks 1.7-mm thick (based on mean thickness of 1.7 ± 0.2 mm (± SE, n = 23) of the encrusting sponges Chondrilla caribensis forma hermatypica, Spirastrella coccinea, S. hartmani, Clathria venosa, Placospongia cf. intermedia, Acarnus nicoleae, and Cliona caribbaea selected at random for measurement in shallow water on the reefs of Belize during May 2019). Sponge volumes were summed by genus within each quadrat to provide a single replicate measure. To measure sponge density and volume in the ~ 5,000 photoquadrats from 1992 to 2017, the 26 year record was analyzed in 2–3 year intervals. All six sites were analyzed for sponge density, but three sites (White Point, Cabritte Horn, and Europa Bay) were selected to estimate sponge volume. The sites for volume estimates were selected to span the range of sponge abundances observed along this shore, and to provide a tractable task commensurate with the limited capacity to estimate organism volume from planar images. Physical environmental conditions As part of the ecological monitoring in this location (Edmunds 2013; Edmunds and Lasker 2016), seawater temperature and rainfall have been recorded since 1989. The records from 1990 to 2017 were used to explore their capacity to account for variation in multivariate sponge assemblage structure, as well as the multivariate community structure defined by sponges, scleractinians, octocorals, macroalgae, and CTB. Seawater temperature was recorded at ~ 9–14 m depth at Yawzi Point using Hobo loggers (± 0.2°C [U22-001, Onset Computer Co., MA]) that sampled every 10–15 minutes. Records were collapsed by day and used to calculate annual summaries of mean temperature, minimum temperature, maximum temperature, the number of hot days (i.e., > 29.3°C), and the number of cold days (i.e., ≤ 26.0°C). From 1992 to 2011, rainfall (cm y-1) was obtained from the Southeastern Regional Climate Center (http://www.sercc.com), which compiled data from a rain gauge in Cruz Bay (Station 671980). Where this record was incomplete, values were obtained from Catherinburg (Station 671348), East End (Station 672551), or a mean value for the missing months calculated from all other values for that same month (in Edmunds and Gray 2014). From 2012, rainfall was measured using a Standard Rain Gauge (NOAA, National Weather Service) deployed on the north shore of St. John (18° 21´ 20.95N, 64° 45´ 57.53W). Temporal variation in the regional-scale physical environment was evaluated through a de-trended index reflecting the effect of the Atlantic Multidecadal Oscillation (AMO) as reported in Kajtar et al. (2019) and provided courtesy of the first author. Hurricane effects were evaluated on a categorical scale assigning impact values to storms of 0 (no storm), 0.5 (minor impacts), and 1 (major impacts), and summing impacts within each year (after Gross and Edmunds 2015). Blank values in this dataset indicate "not applicable." Data analyzed with Systat 14 and PRIMER 6. BCO-DMO Data Manager Processing notes: * Data table extracted from Excel file "Data in Paper 23 July 2020 copy.xlsx" sheet "Fig S2 to csv. * Column removed with heading "METADATA" which describes each of the pervious columns in the sheet. That information was added to the Parameters section of the metadata. * added a conventional header with dataset name, PI name, version date * blank values in this dataset are displayed as "nd" for "no data." nd is the default missing data identifier in the BCO-DMO system. * modified parameter names to conform with BCO-DMO naming conventions (spaces, +, and - changed to underscores). Units in parentheses removed and added to Parameter Description metadata section. |Month||Month in which raw data were collected (four letter month name or three letter abbreviation)||unitless| |Year||Year of survey in format yyyy||unitless| |Site||Site of survey||unitless| |Density||Density of all sponges. Number of sponges per 0.25m^2 quadrat||sponges/0.25 m^2| |Volume||Volume of all sponges. This is volume of the sponges estimated from linear dimensions in the photograms and summed by quadrat. (sponge volume in cm3/ per 0.25 cm2 quadrat)||cm3| Describing how ecosystems like coral reefs are changing is at the forefront of efforts to evaluate the biological consequences of global climate change and ocean acidification. Coral reefs have become the poster child of these efforts. Amid concern that they could become ecologically extinct within a century, describing what has been lost, what is left, and what is at risk, is of paramount importance. This project exploits an unrivalled legacy of information beginning in 1987 to evaluate the form in which reefs will persist, and the extent to which they will be able to resist further onslaughts of environmental challenges. This long-term project continues a 27-year study of Caribbean coral reefs. The diverse data collected will allow the investigators to determine the roles of local and global disturbances in reef degradation. The data will also reveal the structure and function of reefs in a future with more human disturbances, when corals may no longer dominate tropical reefs. The broad societal impacts of this project include advancing understanding of an ecosystem that has long been held emblematic of the beauty, diversity, and delicacy of the biological world. Proposed research will expose new generations of undergraduate and graduate students to natural history and the quantitative assessment of the ways in which our planet is changing. This training will lead to a more profound understanding of contemporary ecology at the same time that it promotes excellence in STEM careers and supports technology infrastructure in the United States. Partnerships will be established between universities and high schools to bring university faculty and students in contact with k-12 educators and their students, allow teachers to carry out research in inspiring coral reef locations, and motivate children to pursue STEM careers. Open access to decades of legacy data will stimulate further research and teaching. Coral reefs are exposed to a diversity of natural and anthropogenic disturbances, and the consequences for ecosystem degradation have been widely publicized. However, the reported changes have been biased towards fishes and stony corals, and for Caribbean reefs, the most notable example of this bias are octocorals ("soft corals"). Although they are abundant and dominate many Caribbean reefs, they are rarely included in studies due to the difficulty of both identifying them and in quantifying their abundances. In some places there is compelling evidence that soft corals have increased in abundance, even while stony corals have become less common. This suggests that soft corals are more resilient than stony corals to the wide diversity of disturbances that have been impacting coral corals. The best coral reefs on which to study these changes are those that have been studied for decades and can provide a decadal context to more recent events, and in this regard the reefs of St. John, US Virgin Islands are unique. Stony corals on the reefs have been studied since 1987, and the soft corals from 2014. This provides unrivalled platform to evaluate patterns of octocoral abundance and recruitment; identify the patterns of change that are occurring on these reefs, and identify the processes responsible for the resilience of octocoral populations. The project will extend soft coral monitoring from 4 years to 8 years, and within this framework will examine the roles of baby corals, and their response to seafloor roughness, seawater flow, and seaweed, in determining the success of soft corals. The work will also assess whether the destructive effects of Hurricanes Irma and Maria have modified the pattern of change. In concert with these efforts the project will be closely integrated with local high schools at which the investigators will host marine biology clubs and provide independent study opportunities for their students and teachers. Unique training opportunities will be provided to undergraduate and graduate students, as well as a postdoctoral researcher, all of whom will study and work in St. John, and the investigators will train coral reef researchers to identify the species of soft corals through a hands-on workshop to be conducted in the Florida Keys. Understanding how changing environmental conditions will affect the community structure of major biomes is the ecological objective defining the 21st century. The holistic effects of these conditions on coral reefs will be studied on shallow reefs within the Virgin Islands National Park in St. John, US Virgin Islands, which is the site of one of the longest-running, long-term studies of coral reef community dynamics in the region. With NSF-LTREB support, the investigators have been studying long-term changes in stony coral communities in this location since 1987, and in 2014 NSF-OCE support was used to build an octocoral "overlay" to this decadal perspective. The present project extends from this unique history, which has been punctuated by the effects of Hurricanes Irma and Maria, to place octocoral synecology in a decadal context, and the investigators exploit a rich suite of legacy data to better understand the present and immediate future of Caribbean coral reefs. This four-year project will advance on two concurrent fronts: first, to extend time-series analyses of octocoral communities from four to eight years to characterize the pattern and pace of change in community structure, and second, to conduct a program of hypothesis-driven experiments focused on octocoral settlement that will uncover the mechanisms allowing octocorals to more effectively colonize substrata than scleractinian corals on present day reefs. Specifically, the investigators will conduct mensurative and manipulative experiments addressing four hypotheses focusing on the roles of: (1) habitat complexity in distinguishing between octocoral and scleractinian recruitment niches, (2) the recruitment niche in mediating post-settlement success, (3) competition in algal turf and macroalgae in determining the success of octocoral and scleractian recruits, and (4) role of octocoral canopies in modulating the flux of particles and larvae to the seafloor beneath. The results of this study will be integrated to evaluate the factors driving higher ecological resilience of octocorals versus scleractinians on present-day Caribbean reefs. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria. |NSF Division of Environmental Biology (NSF DEB)| |NSF Division of Ocean Sciences (NSF OCE)|
Guide Prepared by Michael J. Cummings...© of Work and Year of Publication is a lyric poem in free verse. Sylvia Plath wrote the poem in 1961. The London firm of Faber and Faber Ltd. published it in 1971, eight years after her death, as part of a collection entitled Crossing the Water. The New York firm of Harper & Row published the collection later in the same year. you want to know the truth, be as objective and detached as a mirror. It reflects exactly what it sees without hiding flaws. Whether you are evaluating an actor's performance, a meatloaf recipe, a religion, a political system, a Miss America candidate, a scientific theory, or yourself or another person, you must be "unmisted by love or dislike" (line 3). to the reader, a mirror hanging on a wall says it reflects exactly what it sees. It is not being cruel when it reveals the flaws of the person looking into it, the mirror says. Rather, it is simply being truthful. With proper illumination, it sees everything in front of it. In this respect, it is like a "little god," it says. The mirror says it spends most of its time looking at a pink wall across from it. It is as if the wall has become part of itits heart. From time to time, people pass in front of the wall, making it seem as if the wall is flickering. mirror says it is now a lake. A woman comes by each morning to look into it to examine herself. As she ages, she dislikes what she sees in the lake. But she comforts herself with the insincere flattery of others and the magic of age-banishing cosmetics. (The moon and candles mentioned in line 12 symbolize the false compliments and the cosmetics, for they cast only dim light that does not reveal flaws.) Meanwhile, each time she looks into the lake old age and death rise toward "like a terrible fish" (line 18). Plath wrote the poem in first-person point of view. The speaker is a mirror, which tells the reader what it reflects. In the second stanza, it becomes a lake. For further information see Summary, above. The poem is in free verse, a type of poetry with rhythms based on words patterns rather than meter (such as iambic pentameter). Gustave Kahn (1859-1936) and other French poets pioneered this verse form in the late nineteenth and early twentieth of the Poem remains under copyright. However, a book entitled Poetry Reloaded, by Blair Mahoney, has reproduced the poem with permission of the publisher. You can access the page on which the poem appears by clicking Line 1, I am silver and exact: The word silver here refers to the coating on the back of a glass mirror. It can be made with liquefied silver or aluminum applied to a smooth glass plate. A mixture of silver nitrate and ammonium hydroxide can also be used to make the coating. Line 3, unmisted: Not influenced; not prejudiced. Line 5, eye: The reflecting surface. Line 8, it flickers: The wall alternately disappears and reappears as people pass in front of Line 12, those liars, the candles or the moon: Because candles and moonlight provide only dim illumination, they "lie" about what they see. (See Summary, Stanza 2, for an interpretation of line 12. The most important figure of speech in the poem is personification, in which the mirrorassuming humanlike qualitiesspeaks to the reader about what it reflects. Examples of other figures of speech in the poem include 2: Whatever I see, I swallow Metaphor, line 5: The mirror compares itself to a "little god." Metaphors, line 12: The mirror compares false compliments to the light of the moon and efforts of the woman to hide her signs of aging to the light of candles. Simile, lines 17-18: The mirror compares the image of the aging woman to a "terrible fish." Questions and Writing Topics Can the poem be interpreted as applying to outmoded ideas as well as to an aging appearance? Explain Write a short poem that personifies Sylvia Plath committed suicide when she was only thirty. Is there any evidence that she herself was afraid of aging, like the woman in "Mirror"? (You will need to conduct research to answer this question.) Some aging women and men today frequently undergo facelifts and other procedures to improve their looks. Write an essay evaluating the pros and cons of cosmetic surgery for older
In this section Observation and repeated examinations Why so many examinations? When it comes time for treatment, our goal is to stop the curve from getting worse and to prevent future problems. Not all curves need treatment. Most curves will remain small not get bigger or cause any problems. However we can not predict yet if small curves under 20 degrees are going to progress in the very young. So rather than jumping into unnecessary interventions, our orthopedic specialists may monitor your child if his or her spinal curve is less than 20 degrees. Why does my child need so many X-rays? Finding scoliosis early is very important for successful treatment. Due to the number of repeat examinations we have cutting edge technology to limit your child’s radiation exposure. - EOS: Children's hospital is one of the very few hospitals with an EOS low-dose radiation scanner. When a child has scoliosis, doctors follow his or her growth and development carefully. One of the most important advantages of using the EOS scanner is the large decrease in radiation dose - up to 10 times less than a regular X-ray, and up to 99 percent less than a CT (computed tomography) scan of the spine. In less than 20 seconds, the EOS scanner can take a 3-D, head-to-toe image of your child in a standing, weight-bearing position. EOS images also provide sharper detail, helping providers better decide the right course of treatment and, over time, measure the results of that treatment for children and families living with long-term health conditions like scoliosis. Learn more about the benefits of our EOS scanner, and watch a parent's testimonial video. - 3-D spinal measurement system: Children's is one of the only centers in the country to use the 3-D spinal measurement system, a system that does not use radiation. This system allows us to measure at least 12 different points along the entire length of a child's spine. The 3-D spinal measurement system creates a surface map of the spine so we can measure the change in a child's spinal curve and properly treat his or her scoliosis. It is most commonly used for patients with smaller curves. What kind of therapies can we do? Sometimes physical therapy is included especially if the patient is having pain or the family would like to seek other methods. Physical therapists work very closely with your child's team of orthopedic specialists to ensure your child's curve is closely monitored and care best meets your child's specific needs. We have a very conservative approach to care. We'll work with you and your child so you understand treatment goals. Treatment goals may include sitting with better posture, participation in sports and recreational activities, or simply feeling better about the way he or she looks. There is not good evidence that physical therapy, manipulation or electric stimulation to keep scoliosis size from getting bigger. However, our physical therapists at Children's use special exercises to help manage scoliosis. These exercises are based on your child’s individual curve and activities, whether he or she has pain, and current strength and flexibility. Physical therapy may be a part of your child’s scoliosis treatment plan because it can improve pain, breathing, posture, balance, movement and overall quality of life. Every child's needs and goals are different. Our highly trained physical therapists have found that patients have the best results when we use a variety of treatment methods to manage scoliosis. We use elements of a research-based method called the Scientific Exercise Approach to Scoliosis, as well as other physical therapy methods, to develop a very individualized and practical approach to care. Our goal is to make sure patients develop a greater awareness of their posture and provide them with the tools to maintain improved posture. We do that through: - Flexibility training - Balance and coordination activities - Retraining breathing and posture - Strengthening and conditioning exercises - Therapy and massage techniques Video: The causes and progression of scoliosis Video: Common scoliosis fears of parents and children Request an appointment To make an appointment, call Central Scheduling. Toll free (877) 607-5280
Improving New Zealand's social and economic future by understanding how kids can become better thinkers in digital environments is the aim of a four-year study led by the University of Auckland. A key finding is the need for New Zealand children to be better critical thinkers (reasoning and problem-solving) with better critical literacy, the art of seeing beyond words to the motivations that produced them. Developing in Digital Worlds is being conducted by the Woolf Fisher Research Centre at the University's Faculty of Education and Social Work. It aims to contribute to improved youth education outcomes by accelerating national achievement through digital environments for four to 17-year-olds. Centre director Professor Stuart McNaughton says the study began in 2014 with extensive research and work with a wide range of pre-schools, schools (including the Manaiakalani schools*), students, teachers and families – and is at the forefront of international research on a subject affecting countries around the world. McNaughton says New Zealand stands to benefit socially and economically from the study, funded by the Ministry of Business Innovation and Employment, by applying some of the lessons learned from years of studying kids, as well as teachers and parent. "We've seen that generally New Zealand children have a relatively low rate of critical thinking – one of the key assets people will need, along with critical literacy, in a digital future," he says "We've also seen a low rate of what we call critical reasoning or argumentation – the ability to present an argument but then being able to understand someone else's position and reason out the best result from those various sources." He says many New Zealand students typically take what information is on offer from teachers and then adopt a perspective they rigidly stick to. "That is not the way of the present, let alone the future – it will need thinkers who go deeper than that, who are not only critical thinkers but who have empathy with other people, who are good collaborators with others. That's not only what's required from an economic success point of view but there is plenty of evidence to suggest it is also tied up with our wellbeing." One area of the study involved an online discussion using Google Groups where students debated the environmental issues raised by pop star Taylor Swift filming at Bethells Beach, with particular reference to the effect on protected bird species there. Critical thinking was required to examine conflicting evidence and how information they were exposed to may have had particular agendas, motives and biases – an invaluable skill when sourcing material from the internet. Despite encouragement to search the wider internet for research, most students tended to work with the resources teachers gave them. However, one young girl downloaded a map of the area, demonstrating how far the filming was from the protected birds. This showed, says McNaughton, schools had a tendency to teach persuasive skills (like essays and presentations) but not necessarily critical discussion which take counter-arguments and dissenting evidence into account to reach a better informed conclusion. The study also showed how using digital tools to engage students freed teachers to "do what is quintessentially human about teaching – engaging in dialogue and elaborative exchanges and being able to promote aspects like critical thinking and literacy," says McNaughton. The idea is not to introduce technology for the sake of it, nor alter the curricula or introduce critical thinking periods, but to weave it into the fabric of what teachers do every day to accelerate achievement. McNaughton says there are three main reasons for applying the study's findings and boosting students' critical and analytical abilities: "The first is equity. The digital divide usually refers to lack of access to the tools for some (usually in low decile schools). It can also mean some students simply don't have the same quality of curriculum and teaching resources through and with digital tools. With these results, we can design major interventions to overcome these sorts of disparities. "The second is excellence. Increasing all children's cognitive abilities has an educational and ultimately societal significance; for example, we know the workplace needs people who are good at oblique thinking, problem-solving, collaboration and interaction with others. "But it works in our social space as well; the empathy and consideration for others inherent in thinking which takes other's perspectives into account also plays well into issues like our bad cyber-bullying record and youth mental health issues. There could be a big advantage there." Finally, he says, New Zealand has a real opportunity to innovate in this area as studies like these can be world-leading in identifying and designing better links between teaching, family and digital (including game-based) learning to achieve better outcomes. "We are a small country but we are agile enough to test these ideas out and contribute markedly, not just to our financial capital, but our social capital. If we get this right, we will be able to contribute to children entering society with more advanced cognitive and social skills. "In terms of economic benefit, we know that the cognitive and social skills of the type we are talking about play into national productivity." McNaughton says the research is ongoing while the centre outlines early findings. "We are disseminating a lot of what we have found out now but we're not finished yet because the questions keep changing as you go through the process." *The Manaiakalani schools are a cluster of low decile schools in Auckland who share the vision to lead future- focused learning in connected communities using digital technology.
from plugins import plugin, send_slack_message, ElasticSearchQuery import logging @plugin() class console_logins(object): CLOUDWATCH_RULE_NAME = "sec_alerts_console_login" CLOUDWATCH_FILTER = { "EventPattern": { "detail-type": ["AWS Console Sign In via CloudTrail"] } } def __init__(self, config): self.config = config def match(self, event): if event.get("detail-type") == "AWS Console Sign In via CloudTrail" and event['detail']['eventName'] == "ConsoleLogin" and event['detail']['responseElements'].get('ConsoleLogin') != "Failure": query = ElasticSearchQuery(self.config['ELASTICSEARCH_SERVER'], self.config['ELASTICSEARCH_SERVER_REGION'], "cloudtrail-*") query.add_time_window((60 * 24 * 7)) query.must.append({"term": {"userIdentity.arn": event['detail']['userIdentity']['arn']}}) query.must.append({"term": {"sourceIPAddress": event['detail']['sourceIPAddress']}}) query.must.append({"term": {"eventName": 'ConsoleLogin'}}) query.must.append({"term": {"responseElements.ConsoleLogin": 'Success'}}) query.must_not.append({"exists": {"field": "errorCode"}}) result = query.query(count=True) print result if result: logging.warn("ES query for user console login returned {} results in the last week; not alerting.".format(result)) return False else: logging.warn("ES query for user console login returned {} results in the last week; alerting.".format(result)) return True return False def process(self, event): text = "*{}* (MFA: {}) logged in to the AWS console for *{}* ({}) from `{}` using: ```{}```".format( event['detail']['userIdentity']['arn'], event['detail']['additionalEventData']['MFAUsed'], self.config['ACCOUNTS'][event['account']]['name'], event['account'], event['detail']['sourceIPAddress'], event['detail']['userAgent'] ) message = { "channel": self.config['SLACK_CHANNEL'], "username": "Security-Otter Bot", "icon_url": "http://d.hx.io/1NeN/2oFhQrsA.png", "attachments": [ { "fallback": text, "color": "#36a64f", "title": "AWS Console Login", "text": text, "mrkdwn_in": ["text", "pretext"] } ] } send_slack_message(message, self.config['SLACK_WEBHOOK'])
def solution(A): """ Return the value of the missing element in an array A. :param A: an array of N different integers missing one element :return: the missing element in the array A of different elements """ if not A: return 1 A.sort() if A[0] == 2: return 1 if A[-1] - A[0] == len(A) - 1: return A[-1] + 1 if len(A) == 2: return A[0] + 1 half = int(len(A)/2) first_half = A[:half] last_half = A[half:] if len(first_half) < first_half[-1] - first_half[0] + 1: return solution(first_half) if len(last_half) < last_half[-1] - last_half[0] + 1: return solution(last_half) if first_half[-1] + 1 == last_half[0]: return last_half[-1] + 1 else: return first_half[-1] + 1
Drain wardens (OC22) are placed in drains to capture low level hydrocarbon spills, protecting against low level contamination without attendance. - Drain wardens hold up to 20 litres of oil, diesel or petrol and can be changed when necessary. - They are placed beneath the grate of stormwater inlets, - They act by trapping oil and retaining sediment as contaminated water flows through the device. Oil absorbent bilge socks can also be supplied.
# Dividing the matrices into submarines • N is the size of the matrix, K is the size of the submarine. The matrix is further introduced Accept N = 3, K = 2 Our N*N matrix. ``````1 2 1 1 1 2 1 1 1 `````` All size submarines K*K: ``````[1 2 1 1] [2 1 1 2] [1 1 1 1] [1 2 1 1] `````` How do you do the same job finding a submarine? What do you want me to do? The only thing you've done is create a three-dimensional set of k*k matrices filled with zeros. ``````matrix = [[1, 2, 3, 'a'], [4, 5, 6, 'b'], [7, 8, 9, 'c'], [10, 11, 12, 'd']] size = (3, 2) for j in range(len(matrix) - size[1] + 1): for i in range(len(matrix[j]) - size[0] + 1): for y in range(size[1]): line = [] for x in range(size[0]): line.append(matrix[j + y][i + x]) print(line) print() `````` In a blatant way, you have to go through the matrix so that one element of the submarine is within the range: `matrix_width - submatrix_width``matrix_height - submatrix_height` And then using the coordinates of the first element, find all the other submarine elements in the matrix. or if it is written at the formula level, then: ``````submatrix[x][y] = matrix[i + x][j + y] `````` where (i, j) is the coordinates of the first (left top) submarine element in the matrix P. S. The code can be slightly simplified (for disgusts): ``````[[[print([matrix[j + y][i + x] for x in range(size[0])]) for y in range(size[1])] + [print()] for i in range(len(matrix[j]) - size[0] + 1)] for j in range(len(matrix) - size[1] + 1)] `````` 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
Transparent light purple, transparent greenish blue, and colorless. Method of forming uncertain (see Remarks); finished by cutting, grinding, and polishing. Beaker (?). Rim plain, with flat top, which is ground and polished; wall vertical. Exterior has five, presumably continuous horizontal bands: (from top to bottom) colorless (W. 1.2 cm), purple (W. 0.8 cm), colorless (W. 0.7 cm), blue (W. 0.6 cm), and colorless (W. at least 0.4 cm).. In purple band, outer surface protrudes 0.1 cm beyond colorless surfaces and its inner surface is flush with the inner colorless surfaces; outside haskufic or pseudo-kufic inscription with incised outlines. Blue band has both surfaces flush with surfaces of colorless glass, and is plain. Interior has narrow, horizontal groove immediately above top of purple band.
class Room(object): def __init__(self, name, north, west, east, south, up, down, look): self.name = name self.north = north self.west = west self.east = east self.south = south self.up = up self.down = down self.look = look def move(self, direction): global current_node current_node = globals()[getattr(self, direction)] outside = Room("Outside of House", 'room1', 'shelter', 'garden', None, None, None, "You're outside of a house; there's a garden to the east and a shelter to the west.") garden = Room("Garden", None, None, None, None, None, None, "You're in a garden and there's a shovel.") shelter = Room("Shelter", None, 'inside_shelter', "outside", None, None, None, "You're in a shelter, there's a path to the west and a door to the east.") room1 = Room("Entrance of House", 'room2', None, 'closet', 'outside', None, None, "You entered your home and you have a path to the north and a door to the east and south; there" " is also a table with items.") room2 = Room("Room 2", 'hallway1', 'kitchen', None, 'room1', None, None, "You're in a room and there's a path the the north, west, and south.") closet = Room("Closet", None, 'room1', 'secret_room', None, None, None, "You're in a closet and it's very crowded.") secret_room = Room("Secret Room", None, 'closet', None, None, None, None, "You entered a secret room and you see lots of scrap metal.") kitchen = Room("Kitchen", None, None, 'room2', None, None, None, "You entered a kitchen and there's something on the table.") hallway1 = Room("Hallway 1", None, 'bedroom1', 'bedroom2', 'room2', 'house_f2', None, "You are in a hallway. There's a room to the east, west, and stairs.") bedroom1 = Room("Bedroom 1", None, None, 'hallway1', None, None, None, "You entered a room and there's lots of furniture and items.") bedroom2 = Room("Bedroom 2", None, 'hallway1', None, None, None, None, "You entered a room and there's lots of furniture and items.") inside_shelter = Room("Inside of Shelter", "craft", None, 'shelter', None, None, None, "You entered a room and there is a door to the north and a path to the east.") craft = Room("Crafting Station", None, None, None, 'inside_shelter', None, None, "You are in a room; it seems to be a crafting station.") house_f2 = Room("House F2", None, None, None, None, 'attic', 'hallway1', "You entered in a room and there's a ladder, there's a way down the stairs to the south.") attic = Room("Attic 1", 'attic2', None, None, None, None, "house_f2", "You appeared in an attic; nothing is near you.") attic2 = Room("Attic 2", 'front_house', None, None, 'attic', None, None, "You are in an attic with a window to the window.") front_house = Room("Front of House", "road", None, None, None, None, None, "You appeared on the front side of your house with a road to the north.") road = Room("Endless Road", None, None, None, None, None, None, "You walked endlessly, you've perished" " before you reached your destination.") # END GAME current_node = outside DIRECTIONS = ["north", "east", "south", "west", "up", "down"] short_directions = ["n", "e", "s", "w", "u", "d"] while True: print(current_node.name) print() command = input("CHAT: >_").lower().strip() print("") if command == "look": print(current_node.look) print(" ") if command == "quit": quit(0) # ADD THIS CODE TO NOT REPEAT!!! elif command in short_directions: position = short_directions.index(command) command = DIRECTIONS[position] if command in DIRECTIONS: try: current_node.move(command) except KeyError: print("You cannot go this way.") print()
GeoCOF is a statistical software tool for seasonal forecasting of climatic variables such as rainfall. It facilitates multiple-linear regression modelling between climatic predictors (e.g. sea surface temperatures) and the predictant, usually seasonal rainfall totals. The GeoCOF was developed by the FEWSNET/USGS activity in support of regional seasonal forecasting processes, with assistance from the SADC Climate Services Centre. The GeoCOF rationalizes the "SYSTAT process" with automation of various statistical analyses, improved functions for digitizing ocean basins, and simplified gridding of rain gauge data. Latest GeoCOF download - Current Version (11) - updated August 10, 2016 - Adds an 'about' box and makes some simple fixes on the functionality - Microsoft .NET Framework v4.0. Download it here if you need it. - Windows XP Operating System or Later For more information, contact [email protected]
चिकित्सकों द्वारा प्रमाणित रुप से सिद्ध किया जा चुका है कि जिनको एक से अधिक प्रकार के विभिन्न रोग, जैसे शुगर, फेफड़ों की बीमारी, कैंसर, न्यूरो, हार्ट से संबंधित डिजीज समेत एक या एक से अधिक प्रकार के रोग हैं, जिन्हें चिकित्सकीय भाषा में कोमोरबिलि कोमोरबीडीटी कहा जाता है, उन्हें बचाना बेहद मुश्किल है। वहीं, यह भी पाया गया है कि मधुमेह , हाइपरटेंशन, रक्त संबंधी रोग एचआईवी पॉजिटिव, कैंसर इत्यादि जैसे एक से अधिक रोग से प्रभावित व्यक्ति भी समुचित इलाज एवं देखभाल से कोरोना को मात देकर स्वस्थ हुए हैं। प्रसार की आशंका शारीरिक दूरी बढ़ने के साथ नहीं घटती एसएससी बबलू कुमार ने घटना की पुष्टि करते हुए बताया है कि सूचना मिलते ही पुलिस मौके पर पहुंची और लड़की को उपचार के लिए अस्पताल भर्ती कराया। लड़की के होश में आने पर उसके बयानों के आधार पर आगे कार्रवाई की जाएगी। सड़क एवं परिवहन मंत्रालय ने साल 2019 के दिसंबर में टोल प्लाजा पर फास्टैग अनिवार्य किया था। टोल प्लाजा पर दोनों तरफ एकएक कैश लेन है। बिना फास्टैग लगे वाहनों के कैशलेस लेन में प्रवेश करने पर चालक से निर्धारित दोगुना चार्ज वसूला जाता है। टोल से हर रोज करीब 26 हजार से ज्यादा वाहन</s>
एक बार की बात है, जंगल में एक सप्ताह से बहुत तेज बारिश हो रही थी । इसलिए कोई भी जानवर अपने भोजन के लिए घर से बाहर नहीं जा सका । `हां, आपने ठीक कहा` चूहा बोला । `यदि तुम उस हिरण को पकड़ने में मेरी सहायता करो तो हमें कई दिनों के लिए भोजन मिल जाएगा ।` लोमड़ी ने कहा । लोमड़ी ने सभी को अपनी योजना सुनाई । थोड़ी देर में सभी जानवर अपने-अपने निश्चित स्थान पर योजना के अनुसार खड़े हो गए । लोमड़ी की बात सुनकर चीता बोला, `ओह, यह तो सचमुच बहुत बुरी बात है कि मैं एक चूहे के साथ अपना शिकार बांटू । मैं इस चूहे के साथ खाने की अपेक्षा शिकार को छोड़ना अधिक पसंद करूंगा ।` ऐसा कह कर चीता गुस्से से वहां से चला गया । लोमड़ी की बातों पर विश्वास करके भेड़िया वहां से तुरंत दुम दबा कर भाग खड़ा हुआ । उसने पीछे मुड़ कर भी नहीं देखा । भेड़िये के जाते ही नेवला भी वहां स्नान करके आ गया । वह भी हिरण के मांस का स्वाद चखने के लिए आतुर था । लोमड़ी ने नेवले से कहा, `मैंने चूहे को उठाकर हवा में उछाला, वह जमीन पर गिरते ही मर गया ई मैंने भेड़िए को उसकी पूंछ में पकड़ कर घुमाया और उसके सिर को पत्थर पर दे मारा । पत्थर से टकराते ही उसकी मृत्यु हो गई । मैंने चीते को युद्ध के लिए ललकारा । मुझसे युद्ध हारने के बाद वह अपनी जान बचाने के लिए यहां से भाग गया । अब तुम्हारी बारी है । क्या तुम तैयार हो ?` लोमड़ी की बात सुन कर नेवला डर गया । उसके मुंह से एक भी शब्द नहीं निकला । वह वहां से तुरंत भाग गया । सभी जानवरों के चले जाने के बाद लोमड़ी निश्चिंत होकर आराम से अकेले ही हिरण का मांस खाने लगी ।
# Quick Answer: What Is The Perfect Square Of 28? ## Is 28 a square number? Is 28 a perfect square number. A number is a perfect square (or a square number) if its square root is an integer; that is to say, it is the product of an integer with itself. Here, the square root of 28 is about 5.292. Thus, the square root of 28 is not an integer, and therefore 28 is not a square number.. ## What is the factors of 28? The factors of 28 are the numbers, which on multiplication in pairs results in the original number. These factors are 1,2,4,7,14 and 28. Since, 2 x 14 = 28, therefore (2,14) is a pair factor of number 28. ## What number is the perfect square? The perfect squares are the squares of the whole numbers: 1, 4, 9, 16, 25, 36, 49, 64, 81, 100 … Here are the square roots of all the perfect squares from 1 to 100. 1. ## Is 29 a square number? A number is a perfect square (or a square number) if its square root is an integer; that is to say, it is the product of an integer with itself. … Thus, the square root of 29 is not an integer, and therefore 29 is not a square number. Anyway, 29 is a prime number, and a prime number cannot be a perfect square. ## Is 29 a prime number and why? When a number has more than two factors it is called a composite number. Here are the first few prime numbers: 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 101, 103, 107, 109, 113, 127, 131, 137, 139, 149, 151, 157, 163, 167, 173, 179, 181, 191, 193, 197, 199, etc. ## Is 8 a perfect square? 8 is not a perfect square because you cannot express it as the product of two equal integers. ## Is 2 a perfect square? Answer: YES, 2 is in the list of numbers that are never perfect squares. The number 2 is NOT a perfect square and we can stop here as there is not need to complete the rest of the steps. ## Is 30 a perfect square? Is 30 a perfect square number? A number is a perfect square (or a square number) if its square root is an integer; that is to say, it is the product of an integer with itself. Here, the square root of 30 is about 5.477. Thus, the square root of 30 is not an integer, and therefore 30 is not a square number. ## Is 28 a perfect cube? The value of cube root of one is 28. The nearest previous perfect cube is 27 and the nearest next perfect cube is 64 . Cube root of 28 can be represented as 3√28. ## Why is 28 the perfect number? A number is perfect if all of its factors, including 1 but excluding itself, perfectly add up to the number you began with. 6, for example, is perfect, because its factors — 3, 2, and 1 — all sum up to 6. 28 is perfect too: 14, 7, 4, 2, and 1 add up to 28. ## IS 400 a perfect square? A number is a perfect square (or a square number) if its square root is an integer; that is to say, it is the product of an integer with itself. Here, the square root of 400 is 20. Therefore, the square root of 400 is an integer, and as a consequence 400 is a perfect square. ## What is the perfect square of 29? 841Perfect Square:Positive IntegerInteger Squared=Perfect Squares List2727 ^2 =7292828 ^2 =7842929 ^2 =8413030 ^2 =90092 more rows ## Is 8 a cube number? A cube number is the answer when an integer is multiplied by itself, then multiplied by itself again. The first five cube numbers are: 1, 8, 27, 64 and 125. … ## Is 6 a perfect square? Q: Is 6 a Perfect Square? A: No, the number 6 is not a perfect square. ## How do you know what a perfect square is? Check if given number is perfect squareTake the square root of the number.Take floor/ceil/round of the square root which we got in step 1.Subtract value we got in step 2 from the square root.If the output of step 3 is 0 then the number is perfect square else not.
Worksheets for primary and secondary schools can be downloaded here. - Primary School Student Worksheets (.pdf format) - Junior Secondary School Student Worksheets (.pdf format) - Worksheet for the "Braving the Storm: Hong Kong under Japanese Occupation" Exhibition (.pdf format) - "From Son of Heaven to Commoner: Puyi, the Last Emperor of China" Exhibition Worksheet (.pdf format)
Carolina Digital Literacy seeks to promote advanced communication at UNC-Chapel Hill. Initial efforts will revolve around a pilot stage during the 2016-17 academic year. This site currently supports instructors and students participating in the pilot. These efforts will enable members of the UNC community to: - Communicate using a range of media, both textual and non-textual - Identify the appropriate modes and media for any communication situation - Make decisions about responsible uses of intellectual property when using media - Understand the ethical issues that matter in our contemporary communications landscape - Critically evaluate the messages they encounter in a networked world, assessing their value and situating them within cultural concerns that make a difference to the world. - Operate as not only consumers but also producers of contemporary modes of communication. Equally key, Carolina Digital Literacy will help members of the UNC community to: - Cross intellectual boundaries, recognizing instances when—for instance—approaches from the arts or humanities can engage with sciences or when graphic artists might learn from computer scientists - Work in collaborative teams, relying on the inherent connectivity of contemporary communication to bring people together - Engage with public audiences, connecting knowledge developed through the use of digital tools with local, state, and global communities.
import time high = int(input("请输入塔高")) first = time.time() for line in range(1,high): print(" "*(high-line)+"*"*(line*2-1)) last = time.time() print(last-first)
The question of Triaxial Accelerometers Vs Single axis takes in many factors including cost, accuracy, ease of data collection and more. Below are some facts and illustrations to help you decide which is better for your specific application. Fig.1 Traditional methods for taking data using a single channel data collector is to mount the sensor using a magnet in the X, Y and Z axis. Fig.3 The motivator for fitting triaxial sensors is where portable data collectors are used. Instead of having to prepare three sites on each machine, and take three separate site readings and wait for the settling time at each location, one triaxial sensor alone can achieve all three results. In the right application where access is restricted, triaxial sensors can make for a simple install and save valuable time. Fig.4 Data collectors with 4 channel capability are able to take the signal from each channel of a triaxial accelerometer simultaneously. Since all axis are located at the same point and taken simultaneously, valuable phase information is preserved and can be used in advanced vibration analysis. Fig.5 Vibration data can be different depending on where the vibration is measured. The less direct the path between the exciting element (shaft rotating in bearing) the more attenuated the data. This is more of a problem for high frequency data (gear mesh, roller element bearings) than it is for lower frequency data. Fig.6 In addition if the bearing case is less stiff or has soft feet then vibration amplitudes maybe distorted or amplified. This is caused by the cantilever effect of having the x axis offset from the axis of excitation. Fig.7 For permanently installed systems, the triaxial sensor may provide cost savings for cabling. Installation consists of running only one cable as opposed to three. This maybe significant if the junction box is not located locally. Fig.8 Triaxial sensors are higher cost than three single axis sensors. If local machine topography makes a triaxial attractive then one solution is to mount three single sensors on a triaxial mounting block. As loop powered sensors are never used with data collectors, triaxial 4-20mA sensors are much less relevant. In applications requiring orthogonal mounted sensors a triaxial mounting block can also be used.
Exploring the Benefits of a Constant Current Circuit in Differential Amplifiers • chaoseverlasting In summary, a constant current circuit is added to the emitter of a differential amplifier to increase resistance and reduce common mode gain. chaoseverlasting In a differential amplifier in common mode, a constant current circuit is appended to emmitter part to increase the resistance Re. This results in a very low common signal amplification (which reduces noise in the ckt). Why does a constant current circuit increase the resistance? If you use a real resistor of high resistance, you get way too much voltage drop. Instead, you can use a current sink (or source) circuit to generate the bias current, but still present a high impedance. Then what we're trying to do is control the resultant current, and its not the resistance but the current that causes the common mode gain to be small? chaoseverlasting said: Then what we're trying to do is control the resultant current, and its not the resistance but the current that causes the common mode gain to be small? Not sure I understand the question. The current is the total bias current for the diff pair. The splitting of the current by the diff pair is what gives differential gain, as the current is split unevenly by the difference in input voltages. If the input voltages are the same but varying, the current should not be different in the two sides of the diff pair. chaoseverlasting said: In a differential amplifier in common mode, a constant current circuit is appended to emmitter part to increase the resistance Re. This results in a very low common signal amplification (which reduces noise in the ckt). Why does a constant current circuit increase the resistance? The dynamic resistance. The dynamic resistance of a current source is greater than a resistor. $$R_{dyn} = \frac{dE}{dI}$$ For an ideal current souce, the voltage changes nothing for any change in current. The common mode voltage gain is given by: $$A_c=\frac{\beta R_c}{r_i +2(\beta +1)R_E}$$ (Boylestad, p.600) To minimize this gain, its the resistor $$R_E$$ that we maximize. The bias current also depends on this resistor: $$I_c=\frac{1}{r_i+2(\beta +1)R_E}$$ If we use a constant current circuit to control this current and minimize it, we effectively increase the magnitude of the resistor $$R_E$$ and that causes the common mode gain to be small. Is that right? That is what I meant to say when I said that we use the current to control the gain. Looks good to me, chaos. Finally! What is a constant current circuit? A constant current circuit is a type of electronic circuit that maintains a constant current, regardless of changes in voltage or other factors. This is achieved by using components such as transistors or operational amplifiers to regulate the current flowing through the circuit. What is a differential amplifier? A differential amplifier is a type of electronic amplifier that amplifies the difference between two input signals. This allows for greater precision and accuracy in measuring small changes in the input signals, making it useful in applications such as audio amplification, instrumentation, and signal processing. How does a constant current circuit benefit a differential amplifier? A constant current circuit can improve the performance of a differential amplifier by providing a stable and accurate source of current for the amplifier to work with. This helps to reduce noise and distortion in the output signal, resulting in a cleaner and more accurate amplification of the input signals. What are some potential applications for a constant current circuit in a differential amplifier? Constant current circuits are commonly used in differential amplifiers for various applications such as audio amplification, sensor interface, and precision measurement. They can also be used in communication systems, medical equipment, and industrial control systems. Are there any drawbacks to using a constant current circuit in a differential amplifier? One potential drawback of using a constant current circuit in a differential amplifier is that it may introduce additional complexity and cost to the circuit design. Additionally, if the constant current is not precisely regulated, it may affect the accuracy of the amplifier's output signal. However, these potential drawbacks can be mitigated by careful circuit design and component selection. • Electrical Engineering Replies 1 Views 1K • Electrical Engineering Replies 3 Views 1K • Electrical Engineering Replies 13 Views 1K • Electrical Engineering Replies 21 Views 2K • Electrical Engineering Replies 29 Views 3K • Electrical Engineering Replies 2 Views 1K • Electrical Engineering Replies 24 Views 3K • Electrical Engineering Replies 6 Views 1K • Electrical Engineering Replies 3 Views 3K • Engineering and Comp Sci Homework Help Replies 1 Views 516
Racial Inequity in Traffic Enforcement With the Vision Zero Cities Conference kicking off tomorrow, Transportation Alternatives has released an accompanying collection of essays, the first edition of “The International Journal of Traffic Safety Innovation.” Streetsblog is pleased to republish TA Legislative and Legal Manager Marco Conner’s contribution to the journal. The whole collection is worth your time, and you can download it from TA’s Vision Zero Cities site. As Vision Zero policies are adopted by cities and countries around the world, equity, or a lack thereof, is a major challenge to successful implementation. Equity in Vision Zero is the fair and just implementation of transportation safety measures across all populations, including race, age, gender, geography and socio-economic condition. Where inequities exist in cities, there is also the greatest and most disproportionate rates of traffic deaths and injuries. In U.S. cities, 89 percent of high-income communities have sidewalks, while only 49 percentof low-income communities do. At the same time, black and Latino Americans, who live in low-income communities at higher rates than white Americans, are twice as likely to be killed while walking. These deaths are not accidents, but the result of inadequate and inequitable engineering and transportation policy. They represent the biases that Vision Zero has inherited, and which we must address. There is an urgent need for transportation planners to apply a broad equity analysis to “the three E’s” — engineering, education and enforcement — coupled with policy implementations that are similarly guided. A mandated equity analysis will force engineers, police, and educators to consider, and make an effort to correct, historic wrongs of race, age, gender, geography and socio-economic conditions as they work toward Vision Zero. Today, the most pressing challenge is enforcement as it relates to race in the United States. Here, institutional racism and individual bias against minority groups, particularly black and Latino people, is omnipresent, and traffic enforcement is no exception. A 2015 report in the New York Times found that black drivers across the country are up to five times as likely as white drivers to be stopped and searched while driving, even though police find illegal items less often in black drivers’ vehicles. Black Americans are arrested for drug-related crimes at twice the rate of white Americans, despite white people using drugs at comparable rates and selling drugs at similar or higher rates. From before the point of arrest through adjudication and criminal sentencing, a history of biased, racist and draconian practices have devastated generations and entire communities. In New York City, 12 of the 15 neighborhoods with the most criminal court summonses for bicycle riding on the sidewalk are primarily black or Latino, while 14 of the 15 with the fewest are predominantly white [PDF]. Yet, despite these devastating racial inequities in transportation, enforcement remains a crucial tool in saving lives. The annual number of U.S. drunk driving deaths decreased from 25,000 in 1980 to fewer than 10,000 today due to increasing the frequency and certainty of enforcement. Enforcement by automated speed cameras in New York City has changed driver behavior significantly enough to reduce speeding by up to 60 percent [PDF]. Individuals and institutions dedicated to Vision Zero cannot summarily dismiss enforcement as a tool, or ignore systemic injustices. There is hope, however, in a nuanced approach to enforcement, where areas like automated cameras, data-driven safety summonsing, or the emergency response to a traffic-injury crash do not exhibit racial bias. Here, cities can find a framework for increasing certain traffic enforcement measures, and by applying comprehensive institutional changes, overcome existing inequities. Enforcement should be considered in three major categories: traffic-injury crashes, traffic safety violations, and investigatory traffic stops. Additionally, there are two stages to all enforcement: first, the initial police response up to, but not including, arrest or ticketing of the driver; and second, the point of arrest or ticketing, including prosecution, adjudication, and penalties. In a traffic-injury crash, a victim has been harmed, frequently by a driver who made a conscious dangerous driving choice, and police take action in response. For drivers involved in causing these crashes, little or no racial disparity exists within the first stage of enforcement. However, bias does occur in the second stage of enforcement, with black and Latino people far more likely to be arrested, receive less favorable plea offers, and harsher punishments than similarly situated white people, across all types of crime. Traffic Safety Violations For a driver caught committing a traffic safety violation, like speeding or failure to yield, little racial disparity is seen in summonsing by police in either stage of enforcement. Because moving violations and their penalties are clearly defined there is little room for subjectivity. “The driver’s race, age and gender has no relevance to the likelihood of being stopped,” write Charles Epp, Donald Haider-Markel, and Steven Maynard-Moody in Pulled Over: How Police Stops Define Race and Citizenship. Rather, it is driver behavior that determines police enforcement of most traffic safety laws. Using automated camera enforcement wholly removes racial bias. Investigatory Traffic Stops By contrast, investigatory traffic stops see the most blatant racial disparities, from before arrest or ticketing through adjudication and sentencing. Investigatory stops are typically discretionary police stops for strikingly minor violations like driving too slowly or failing to signal a turn. “Virtually all of the wide racial disparity,” the researchers note in Pulled Over, “is concentrated in one category of stops: discretionary stops for minor violations of the law.” Investigatory traffic stops mimic stop-and-frisk tactics common in major U.S. cities, part of a policy of so-called “broken windows” policing, where minor offenses are targeted under the theory that this prevents major crimes. Indeed, these traffic stops have become known as “broken taillight policing,” where a violation relatively insignificant to safety is aggressively and subjectively enforced. The results are the disparate stopping, ticketing and arresting of drivers and bicyclists in predominantly African-American neighborhoods. Broken taillight policing criminalizes nonviolent and non-criminal behavior, and thus risks creating opposition to enforcement against dangerous driving. Further, because the summonses and arrests that result are tried in a racist criminal justice system, investigatory traffic stops are inherently inequitable. Without addressing this bias in enforcement, we will be unable to reach Vision Zero, because all-important traffic enforcement resources will be misdirected, perpetuating linked cycles of racial bias and ineffective traffic enforcement. Today, remarkable technology and a deep pool of data allow cities to reduce traffic fatalities while avoiding the biases that exist in their police systems. Data-driven enforcement, which targets the most dangerous driver behaviors in the locations where they are causing the most harm, is a focused road map for law enforcement. Automated camera enforcement effectively protects against racism, removing the subjectivity of police officers, which is common in investigatory traffic stops. Cities that wish to reach Vision Zero are also responsible for helping enforcement agencies reach a less biased horizon. Educating officers on the moral significance of focusing on the most dangerous driver behaviors, providing anti-racism training and education, and reaffirming that a specific and articulable danger connected to a specific driver is the only standard for a traffic stop are all necessary first steps. To help address racial disparities following arrest, minor offenses like bicycling on the sidewalk should be decriminalized. Additionally, police, prosecutors and courts must be fully transparent and provide aggregate demographic and case data to help accurately define problem areas and measure progress. To clear the slate, and set a strategic plan for easing inequities, cities should commission third-party assessments on the role that race plays in their enforcement and establish community-based restorative justice programs. Finally, police departments must have meaningful independent civilian oversight to create accountability and help overcome decades of institutional racism. There is no acceptable form of Vision Zero without these steps. For all cities adopting Vision Zero, an intersectional and inclusionary equity analysis must permanently guide engineering, education and enforcement along the lines of age, gender, geography and socio-economic condition as well as race. Equity must become a fourth “E,” applied in a recurring process of analysis, implementation, and evaluation. Achieving equity in Vision Zero is not only a moral obligation; equity is a tool and tactic requisite to reach our goal.
अंत में राज्य के अनेक हिस्सों में मानसूनपूर्व वर्षा होने की संभावना है। गायेशपुर शहर में रविवार को बिजोय सिल नामक 34 एक व्यक्ति अपने घर के समीप एक पेड़ से फांसी के फंदे से लटका मिला था। हिंदी न्यूज़ उत्तर प्रदेश बदायूं आशा नहीं काम की, कर दो बर्खास्त किसी भी सेंटर से कर सकते है प्राप्त इलेक्ट्रॉनिक उपकरण से नकल करते युवक पकड़ा उप जिलाधिकारी सौरभ कुमार सिंह, सीओ श्रीयश त्रिपाठी के नेतृत्व में कोतवाली पुलिस व पीएसी के जवानों ने शहर के सिविल लाइंस, सुभाष चौक, मालवीय रोड, नगर पालिका रोड, स्टेशन रोड में फ्लैग मार्च किया। इस दौरान पूजा समितियों के आयोजकों से मुलाकात कर डीजे नहीं बजाने का निर्देश दिया। एपवा ने किया प्रदर्शन राजेंद्र चौधरी ने बताया कि संयुक्त प्रेस कान्फ्रेंस रविवार को एक बजे होटल ताज के विवांता में होगी। इसी मौके पर फिल्म सुल्तान के गीत की तर्ज पर नया नारा यूपी को साथ पसंद है जारी किया जाएगा। कांग्रेस छोड़कर झामुमो में आए थे बस बात इसी बात पर अटकी है कि ये फिल्मों दर्शकों को कितना लुभा पाएंगी। पिछले साल बॉक्स ऑफिस की सबसे दमदार भिड़ंत में जहां पहले हफ्ते में शाहरुख खान की दिलवाले ने बढ़त बनाई थी लेकिन</s>
''' Multiple plots on single axis It is time now to put together some of what you have learned and combine line plots on a common set of axes. The data set here comes from records of undergraduate degrees awarded to women in a variety of fields from 1970 to 2011. You can compare trends in degrees most easily by viewing two curves on the same set of axes. Here, three NumPy arrays have been pre-loaded for you: year (enumerating years from 1970 to 2011 inclusive), physical_sciences (representing the percentage of Physical Sciences degrees awarded to women each in corresponding year), and computer_science (representing the percentage of Computer Science degrees awarded to women in each corresponding year). You will issue two plt.plot() commands to draw line plots of different colors on the same set of axes. Here, year represents the x-axis, while physical_sciences and computer_science are the y-axes. ''' # Import matplotlib.pyplot import matplotlib.pyplot as plt # Plot in blue the % of degrees awarded to women in the Physical Sciences plt.plot(year, physical_sciences, color='blue') # Plot in red the % of degrees awarded to women in Computer Science plt.plot(year, computer_science, color='red') # Display the plot plt.show()
Send the link below via email or IMCopy Present to your audienceStart remote presentation - Invited audience members will follow you as you navigate and present - People invited to a presentation do not need a Prezi account - This link expires 10 minutes after you close the presentation - A maximum of 30 users can follow your presentation - Learn more about this feature in our knowledge base article Do you really want to delete this prezi? Neither you, nor the coeditors you shared it with will be able to recover it again. Make your likes visible on Facebook? You can change this under Settings & Account at any time. Copy of RADIO PLAY Transcript of Copy of RADIO PLAY WORDS In the use of dramatic content, words are possibly the most important element of creating a radio drama, for without words what is ascript.In radio drama, words are used within the script for different purposes such as:To direct and instruct:Using words within a script allows you to directthe actors how to perform certain lines andeven when and where to deliver them.To Produce Dialogue:Without dialogue a radio drama is useless, as diaglogue is a key feature to telling the story,and the actors will need to see a script with words in order to know what their going to say and what the story is. Effects can be the cherry on top of your radio drama cake, as they provide that very necessary touch of realism to your radio drama. Effects are used to intensify the setting of the scene.You can use sound effects to really set the scene of anything, for example:Using a police car siren and, breaking glass, gunshots and helicopter fans for a bank robbery. Don't forget you can also use voice to create sound effects too not just objects and instruments. Aural Signposting are:• mowing the lawn• dog barking• knock on• old-fashioned shop door bell• door. Aural signposting may fail if not used correctly, as it may not set the correct scenery. What Is Radio Play? A dramatized audio performance, broadcast on radio or published on audio media, such as tape or CD. Because there is no accompanying video, a radio drama depends on narration, dialogue, music and sound effects to help the listener imagine the characters and story. History of Radio Play By the late 1930s, radio drama was widely popular in the United States (and also in other parts of the world).There were dozens of programs in many different genres, from mysteries and thrillers, to soap operas and comedies programs in many different genres, from mysteries and thrillers, to soap operas and comedies. Music is one feature used within radio drama that can change the theme,setting, and even storyline of your radio drama, as music sets the scene and mood of the story (music does this in visually interactive entertainment systems such as TV, and it's intensified with radio as picture cannot be included to change the scene or the mood).You can use music in your radio drama by using jingles or intervals between each scene, which could help you to change the location or time of your radio drama, for example: By using music you can paint the picture of day or night to your audience by using instruments like violins or guitars to separate the two scenes. Voices can be used by your actors to help differentiate your characters, as your actors can alter their voice, accent or volume to make spereractions between the characters, which is extremely important as in radio drama the audience cannot see the different characters so they will need to hear the different charaters. For example, this could be done to separate two characters that may seem similar in physical description, as the wealthier character would talk differently (more posh, well spoken and sound better educated). Flashback is a change in a narrative to an earlier event that disturbs the normals sequential development of a story. Flashbacks usually occurs when a character tells a story or has a dream. Flashbacks are an excellent method to putting an exciting twist to the average routine of beginning, middle and end, their also a good method to revealing more about your character's motivations and influences in a short manner. In radio drama,flashbacks are an exciting way to let your audience know more about your characters and develop a relationship and opinion of them, without giving away too much.If not done well flashbacks could became too complex for your audience and confuse them. A cliff-hanger ending is when the storyline of any film, episode, book or radio drama etc. end without resolution. A cliff-hanger ending usually leaves the characters in a dramatic situation, in which the audience normally don't see resolving or overcoming. Cliff-hanger endings are an excellent method to maintaining or extending a high level of suspense and inconclusive thoughts to your audience. Cliff-hanger endings are also great to use for radio drama for a series, as they leave the audience wanting more (just like in a television series).The effect of Cliff-hanger endings could back fire if used to often as the audience will get fed up and irritated by the high level of suspense. Narration is the telling of a story. Any story, either told from 3rd person, 1st or 2nd. It is mostly telling a story, of events that have happened. It can also mean description,explanation, reading, recital, storytelling, voice-over.Narration is a very good device to do a lot of things, as through voice overs you either.• Be able to describe the story.• Be able to voice the characters thoughts• Be able to give your story a more interesting twist. Use of silence Silence in radio drama can be one of the most effective yet simple convention of your radio drama. Silence is used in a radio drama for a lot of reasons and can be two different kinds of silence e.g.Complete Silence –Used as a way to change a scene or emphasis the end of any kind of dialogue or action. Silence within the scene used as a way to emphasize a dramatic or awkward atmosphere. If silence is used too often or even not enough it could seem completely unnecessary and make your radio drama seem flawed or even incomplete. Characterisation is the art of creating characters for a narrative, including the process of conveying information about them.This could involve information about their personality, physical appearance or background story e.g. Linda Smith was a very curious young girl and shy young girl, who constantly wore a flowing flowery dress and rose within her hair. The duration of your radio drama is a very important factor to take into account when creating a radio drama, as radio is not a device that most individuals have the patients to listen to for too long. Listeners can get bored or annoyed if the radio drama last too long,and if listeners have tuned into the radio drama while participating in other task (as majority of listeners do) such as,driving or working, they will tune out as soon as they have completed the task,whether the radio drama is finished or not. A typical duration of a radio drama is anywhere between (3-10 minuets), the lengthy radio drama‟s vary between (25-30 minuets) Its important to structure your storyline, in an entertaining yet simple manner that the audience will not get confused yet stay intrigued with your radio drama. Unlike a film or television program, the audience cannot physically see the story, making it more challenging to follow the storyline, so listeners who tune in, later within the story get easily lost the story (also easily bored). Development of the plot is extremely important, if the storyline of your radio drama doesn't develop in interest value,listeners will get bored and tune out.Your radio drama must have an interesting start and a constantly peaking middle and the climatic end.
इस तरह के नोट में किसी का हस्ताक्षर नहीं होता और आम तौर पर एक संस्थान दूसरे संस्थान को ऐसे नोट सौंपते हैं. ग्रुप का सम्मान करते हुए छोड़ी सीट : इसके बाद माथुर स्टेडियम में स्वामी रामदेव का लोगों ने अभिनंदन किया और शहर के विभिन्न मार्गो से होते हुए स्वागत जुलूस निकाला. राजस्थान प्रदेश कांग्रेस कमेटी की उपाध्यक्ष एवं मीडिया चेयरपर्सन डॉ. अर्चना शर्मा ने नगर निगम द्वारा आम नागरिकों के साथ... पढ़ें पदेन सदस्यों में बंगाल के लिए एआईसीसी प्रभारी जितिन प्रसाद, प्रदेश कांग्रेस प्रमुख अधीर रंजन चौधरी और कांग्रेस विधायक दल के नेता अब्दुल मन्नान के अलावा पश्चिम बंगाल के लिए एआईसीसी सचिव भी शामिल हैं। मुख्यमंत्री ने इन नवनियुक्त पदाधिकारियों से कहा कि वे अपनीअपनी भूमिका में कड़ी मेहनत कर आमजन की सेवा करें और राज्य के विकास में अपना योगदान दें। उन्होंने कहा कि इसी उद्देश्य के लिए उनका चयन किया गया है। उन्होंने लोगों से मुख्यमंत्री जल स्वावलंबन अभियान में भागीदारी निभाने और विशेष पिछड़ा वर्ग के लिए आरक्षण पर राजस्थान हाईकोर्ट के निर्णय के बाद सामाजिक सद्भाव बनाए रखने की अपील की। न्यूली मैरिड कपल पुनीत और निधि ने सोशल मीडिया पर अपनी वेडिंग फंक्शन की कई तस्वीरें और वीडियो शेयर किए</s>
import sqlite3 with sqlite3.connect("new.db") as connection: #SQL functions c = connection.cursor() sql = {'average': "SELECT avg(population) FROM population", 'maximum': "SELECT max(population) FROM population", 'minimum': "SELECT min(population) FROM population", 'sum': "SELECT sum(population) FROM population", 'count': "SELECT count(city) FROM population"} for keys, values in sql.items(): #execute each query c.execute(values) # fetchone() retrieves one record from the query result = c.fetchone() print (keys + ":",result[0]) """ 1. Essentially, we created a dictionary of SQL statements and then looped through the SQL Functions dictionary, executing each statement. 2. Next, using a for loop, we printed the results of each SQL query. """
HEPATITIS B IN IMMIGRANT POPULATIONS: BREAKING BARRIERS AND IMPROVING OUTCOMES Ahmet Gurakar, MD* First discovered in 1965, hepatitis B is a double-stranded DNA virus that is known to infect 350 to 400 million people worldwide, generating 50 million new cases and resulting in 520 000 deaths annually.1-3 Although the rate of infection with the hepatitis B virus (HBV) has decreased significantly as a result of successful vaccination programs, chronic HBV remains the leading cause of cirrhosis and hepatocellular carcinoma (HCC) in endemic areas, including much of Europe, Asia, Africa, the Middle East, and the Pacific Islands, where the regional prevalence of hepatitis B surface antigen (HBsAg) positivity ranges from 2% to 8%.4 In general, the frequency of HBV infection and patterns of HBV transmission vary dramatically in different parts of the world. Approximately 45% of the world’s population lives in areas where the prevalence of chronic HBV infection is high (>8% of the population is HBsAg-positive), 43% live in areas of intermediate prevalence (2%–7% of the population is HBsAg-positive), and only 12% of the world’s population lives in areas of low endemicity (<2% of the population is HBsAg-positive). In highly endemic regions, the lifetime risk of HBV infection is more than 60%.5 These areas include most of Asia (except Japan and India), most of the Middle East, the Amazon Basin of South America, most Pacific Island Groups, Africa, and regions with certain populations (eg, Native Alaskans, Australian Aborigines, and Maoris in New Zealand).5 In these parts of the world, most infections occur at birth (perinatally) or during early childhood, when the risk of acquiring chronic infection is the greatest. There is little recognition of acute disease because most early childhood HBV infections are asymptomatic. In areas of intermediate endemicity, the lifetime risk of HBV infection ranges from 20% to 60% and infections occur in all age groups. Acute disease is commonly recognized because many infections occur in adolescents and young adults. In areas of low endemicity (eg, the United States), the lifetime risk of infection is less than 20% and most infections occur among adults in well-defined risk groups (eg, injection drug users and men who have sex with other men).5 In the United States, a marked decrease in the prevalence of chronic HBV infection among younger, vaccinated, foreign-born US residents is contrasted by an increasing rate of liver cancer deaths among certain foreign-born populations, particularly Asian/Pacific Islanders. Clearly, the prevalence of HBsAg (indicator of chronic HBV infection) in the United States appears to be impacted by migratory trends from regions of high endemicity. From 1996 (when the data were gathered for the HBsAg prevalence) to 2002, approximately 1.9 million, 933 000, 337 000, and 401 000 persons immigrated from Asia, Europe, Africa, and South America, respectively. 6 The prevalence of HBV among Asian Americans ranges from 10% to 15%, and HCC ranks as the first, second, and fourth leading cause of death among Asian American men living in California, Vietnamese and Cambodian Americans, and Chinese and Korean Americans, respectively.7 Among persons immigrating to the United States from the Middle East and Africa, the prevalence of chronic HBV is 5% to 15%, and is higher in men than in women.4 Alarmingly, 25% of these immigrants are women are of reproductive age, a factor that increases the risk of perinatal transmission of HBV. The prevalence of chronic HBV among immigrants from Eastern European counties (except Hungary) is 2% to 7%.4 In the past, the Centers for Disease Control and Prevention recommended screening only persons born in geographic regions with HBV endemicity of 8% or higher (Figure 1), but the agency recently updated screening recommendations to include all persons born in geographic regions with HBV prevalence of 2% or higher (Figure 1).4,8 RISK OF VERTICAL TRANSMISSION Hepatitis B virus is present in body fluids, including blood, saliva, sweat, breast milk, tears, urine, vaginal secretions, semen, and menstrual blood.9 Viral transmission may occur via mother-to-child (vertical or perinatal transmission) or percutaneous or mucosal exposure to infectious bodily fluids (horizontal transmission). In areas with a low prevalence of HBV, horizontal transmission (eg, sexual contact, injection drug use, or occupational exposure to blood or blood products) is more common. In regions of high endemicity, which is particularly relevant to this discussion, perinatal transmission remains the predominant mode of HBV acquisition. If a pregnant woman contracts acute HBV infection during the first or second trimester, HBV rarely infects the infant. However, if infection occurs in the third trimester or in the postpartum period, it will likely lead to infection in the infant.1 Infants born to mothers who are both hepatitis B envelope antigen (HBeAg) (+) and HBsAg (+) have a 70% to 90% risk of developing chronic HBV infection by 6 months of age (in the absence of postexposure prophylaxis), whereas those born to HBeAg (-) and HBsAg (+) mothers have a considerably lower risk of infection (10%).10 Because immediate postnatal vaccination with both passive and active immunoprophylaxis of at-risk infants born to HBsAg (+) mothers significantly reduces the risk of transmission, current guidelines recommend that infants born to infected mothers receive the HBV vaccine and HBV immune globulin within 12 hours of birth. Infants born to mothers whose HBsAg status is unknown should receive the HBV vaccine within 12 hours of birth and the mother should have blood drawn immediately. If the mother is determined to be HBsAg (+), her infant should receive HBV immune globulin at no later than 1 week of age.10 HBV DISEASE PROGRESSION As illustrated in Figure 2, progression of HBV infection is variable. Acute HBV infection is usually selflimiting and benign, but in a proportion of patients, it may progress to chronic HBV infection, leading to more serious complications, including cirrhosis, liver failure, and HCC.11-14 Initially, acute infection (shortterm infection) causes the body to mount an immune response, which essentially involves eradicating the virus by destroying HBV-infected liver cells. In some patients (ie, adults or adolescents), the immune response succeeds and the patient experiences a complete recovery (resolution). In other HBV-infected individuals (ie, infants), however, the immune response is insufficient in eliminating the virus acutely, and thus the immune system perseveres and may eventually (over months, many years, or indefinitely) lead to mass destruction of infected liver cells as the virus spreads (ie, long-term infection). This slow, persistent destruction of liver cells by the immune system can lead to fibrosis, cirrhosis, and HCC. Patients infected with HBV whose immune systems cannot eliminate the virus are referred to as chronic HBV carriers, and these individuals can transmit the virus to others. The reason why chronic infection cannot be cured may be related to the ability of the HBV virus to enter the nucleus of liver cells and produce covalently closed circular (ccc) DNA, which is essentially a permanent blueprint for the virus. Each time the virus reproduces, it replenishes the supply of cccDNA and, although currently available treatments can reduce HBV DNA to undetectable levels, they cannot eliminate the cccDNA. Hepatitis B virus-related morbidity and mortality is alarming and, at times, difficult to detect early, as evidenced by cases in which HCC developed in relatively young patients and in the absence of any previous liver disease. In those who develop cirrhosis, the 5-year life expectancy drops substantially. One study found the 5-year survival in 130 patients with cirrhosis (27% of whom had ascites; mean serum bilirubin was 1.6 mg/dL and mean albumin 3.5 g/dL) was 55%.15 In another study of 21 patients with cirrhosis and hepatic decompensation (determined by presence of ascites, jaundice, encephalopathy, and/or a history of variceal bleeding) the 5-year survival was 14%.16 The relationship between chronic HBV infection and HCC has become increasingly clear in recent years, with one major study correlating serum HBV DNA levels with the risk of developing the malignancy.17 In this prospective trial of 3653 HBsAg (+) patients (aged 30–65 years) in Taiwan, the incidence of HCC increased with serum HBV DNA levels in a dose-response relationship, ranging from a 1.3% rate for an HBV DNA level of less than 300 copies/mL to a 14.9% rate for an HBV DNA level of 1 million copies/mL or more. Participants with persistent elevations in serum HBV DNA levels during follow- up had the highest risk of HCC. This study underscores both the importance of determining viral load and utilizing current treatment regimens to achieve undetectable viral levels. This issue of Johns Hopkins Advanced Studies in Medicine is based on the proceedings of a national series of “lunch and learn” meetings, which were held from August to October 2009. This monograph is dedicated to educating primary care physicians, nurses, nurse practitioners, and physician assistants on screening for and managing chronic HBV infection in at-risk immigrant populations within the United States. Tarek Hassanein, MD, explores the life cycle of HBV as it relates to perinatal or early childhood- acquired infection, and reviews the various phases of HBV infection. He also discusses diagnostic and screening strategies (including the recently updated recommendations from Morbidity and Mortality Weekly Report), along with public health implications of treating HBsAg (+) individuals. In discussing management of chronic HBV infection, Norman L. Sussman, MD, focuses on US Food and Drug Administration-approved regimens, treatment duration and end points, and therapeutic monitoring. He also addresses management of chronic HBV infection in special populations, specifically in pregnant patients and in those undergoing chemotherapy for solid organ or hematologic malignancies. 1. Lok AS, McMahon BJ. Chronic hepatitis B. Hepatology. 2007;45:507-539. 2. Hoofnagle JH, Doo E, Liang TJ, et al. Management of hepatitis B: summary of a clinical research workshop. Hepatology. 2007;45:1056-1075. 3. Hu KQ. Hepatitis B virus (HBV) infection in Asian and Pacific Islander Americans (APIAs): how can we do better for this special population? Am J Gastroenterol. 2008;103:1824-1833. 4. Weinbaum CM, Williams I, Mast EE, et al. Recommendations for identification and public health management of persons with chronic hepatitis B virus infection. MMWR Recomm Rep. 2008;57(RR-8):1-20. 5. World Health Organization. Geographic prevalence of HBsAg. Available at: http://www.who.int/vaccines-surveillance/ graphics/htmls/hepbprev.htm. Accessed July 8, 2009. 6. US Citizenship and Immigration Services. 2002 Yearbook of Immigration Statistics. US Department of Homeland Security Web site. Available at: http://uscis.gov/graphics/shared/ aboutus/statistics/yearbook/2002.pdf. Accessed July 8, 2009. 7. Asian Liver Center. Stanford University School of Medicine Web site. Available at: http://liver.stanford.edu. Accessed June 4, 2009. 8. Traveler’s Health—Yellow Book. Centers for Disease Control and Prevention. Available at: http://wwwn.cdc.gov/travel/ yellowbook/ch4/hep-b.aspx. Accessed June 5, 2009. 9. Boag F. Hepatitis B: heterosexual transmission and vaccination strategies. Int J STD AIDS. 1991;2:318-324. 10. Centers for Disease Control and Prevention. A comprehensive immunization strategy to eliminate transmission of hepatitis B virus infection in the United States: recommendations of the Advisory Committee on Immunization Practices (ACIP); Part 1: Immunization of Infants, Children, and Adolescents. MMWR Morb Mortal Wkly Rep. 2005;54(RR-16):1-31. 11. Torresi J, Locarnini S. Antiviral chemotherapy for the treatment of hepatitis B virus infections. Gastroenterology. 2000;118(2, suppl 1):S83-S103. 12. Moyer LA, Mast EE. Hepatitis B: virology, epidemiology, disease, and prevention, and an overview of viral hepatitis. Am J Prev Med. 1994;10(suppl):45-55. 13. Fattovich G, Giustina G, Schalm SW, et al. Occurrence of hepatocellular carcinoma and decompensation in western European patients with cirrhosis type B. The EUROHEP Study Group on Hepatitis B Virus and Cirrhosis. Hepatology. 1995;1:77-823. 14. Perrillo RP, Wright T, Rakela J, et al. A multicenter United States-Canadian trial to assess lamivudine monotherapy before and after liver transplantation for chronic hepatitis B. Hepatology. 2001;33:424-432. 15. Weissberg JI, Andres LL, Smith CI, et al. Survival in chronic hepatitis B: an analysis of 379 patients. Ann Intern Med. 1984;101:613-616. 16. De Jongh FE, Janssen HL, de Man RA, et al. Survival and prognostic indicators in hepatitis B surface antigen-positive cirrhosis of the liver. Gastroenterology. 1992;103:1630-1635. 17. Chen CJ, Yang HI, Su J, et al. Risk of hepatocellular carcinoma across a biological gradient of serum hepatitis B virus DNA level. JAMA. 2006;295:65-73. *Associate Professor of Medicine, Division of Gastroenterology and Hepatology, Medical Director, Liver Transplantation, The Johns Hopkins University School of Medicine, Baltimore, Maryland. Address correspondence to: Ahmet Gurakar, MD, Associate Professor of Medicine, Division of Gastroenterology and Hepatology, Medical Director, Liver Transplantation, The Johns Hopkins University School of Medicine, 1830 East Monument Street, Suite # 429, Baltimore, MD 21205. E-mail: [email protected]. The content in this monograph was developed with the assistance of a medical writer. The authors made substantial contributions to the intellectual content of the articles by conceiving and designing the original presentations, researching references and studies, drafting the manuscripts, reviewing and revising the manuscripts, and/or providing supervision.
चीन इफोन क्स के लिए उच्च गुणवत्ता वाला रंगीन बैक केस निर्माता होम > उत्पादों > फ़ोन मामले > इफोन क्स के लिए उच्च गुणवत्ता वाला रंगीन बैक केस इस स्टाइलिश और सुरुचिपूर्ण केस कवर में स्पीकर, चार्जिंग पोर्ट और बटन के लिए सटीक कटआउट हैं। बटन मूल क्लिक प्रतिक्रिया और त्वरित पहुँच प्रदान कर सकते हैं। यह केस ९ह टेम्पर्ड ग्लास बैक कवर , एक परफेक्ट बम्पर और शॉक-एब्सॉर्बेंट केस सहित बेहद स्क्रैच-रेसिस्टेंट है, जो आपके इफोन क्स के बैक बॉडी और किनारों को डेली ड्रॉप्स और बंप्स से बचाता है। --- ए: यदि हमारी ओर से कोई समस्या होती है, तो हम निश्चित रूप से ग्राहकों को इन समस्याओं को हल करने में मदद करेंगे। स्ट्रैप वाला फोन केस
There are probably going to be several engines around a typical lease pumping operation. Obviously, there’s the engine in the truck that gets you out to the lease site. There may also be generators if the lease is far from electrical lines, or as a backup power source. They can power pumps, and do a number of other vital jobs. The engines on the lease may run constantly, and so the ability to maintain and repair the engines around the site is an important one. This is particularly true with smaller and wildcat operations, when even small expenses can have an impact on profitability. Just about every modern engine is an internal combustion engine, meaning it’s powered by small, constant explosions. Gas and air are mixed within the engine, in the cylinder, and then compressed by a piston. With some engines, a small spark ignites the mixture. In diesel engines, the compression is enough to cause the explosion. The piston is driven down by the explosion, which turns a crank and provides power. In four cycle engines, the piston moves four times per explosion. There are also two cycle engines which moves two times per explosion, but four cycle engines are more common. The first downstroke is the intake stroke, during which air and fuel are drawn into the cylinder through the intake valve. When the piston approaches the bottom of the cylinder, the intake valves open. In the first upstroke, called the compression stroke, the fuel and air mixture is forced upward and compressed into 1/10 the original volume. This heats the mixture. The compressed mixture is ignited by the spark plug and the piston is forced down in the power stroke by the explosion. The cylinder is now filled with the byproducts of the explosion, which are forced out in upward exhaust stroke through the exhaust valve. Each step must happen at the correct moment for the engine to be as efficient as possible. Engine timing is an important part of how the engine works, and a mis-timed engine will often cease to work altogether. Figure 1. Single cylinder engines are common around oilfields. This one has a large flywheel to provide smooth power. (courtesy Arrow Specialty Co., Inc.) A large number of the engines around the operation will be single cylinder engines, meaning they’ll have only one cylinder and piston. Circulating pumps often are powered by single, vertical cylinder engines, and larger horizontal, cylinder engines are used to power mechanical pumps for some wells. Single cylinder engines that operate pumps usually are slower and have large flywheels. The flywheels use momentum to even out the power generated by the single cylinder engine. For the single cylinder engine, everything is kept relatively simple. There’s a single spark plug and a simplified electrical system. The sheave, which attaches to a belt to provide power to the pump or other piece of equipment, is usually larger and revolves at a slower speed. Figure 2. To the left of the picture you can see a single cylinder engine using gas produced from the well as fuel. Engines with multiple cylinders have correspondingly more complex requirements when it comes to electrical and fuel systems, and may need a separate battery, voltage regulator, coil and more for each cylinder. Multi-cylinder engines also usually have a smaller flywheel and so run much more quickly. Maintaining a working engine is always preferably to repairing a broken one. The ability to performance basic, or even more complex, maintenance on the engines around your operation is a useful skill. Some things require the services of a shop with specialized equipment and specific know-how. Most problems having to do with the piston, cylinder, or valves are going to fall into this category. Most regular maintenance and some basic repairs can be done with a basic understanding of how each system works. The electrical system in your car powers the radio, and probably the locks and windows. For most engines around the lease, however, the electrical system is used just to provide the spark which ignites the fuel in the cylinder. The spark is provided by the spark plug. Spark plugs must be the right size and within the right temperature range. These can include standard, hot, and cold. The points on the spark plug must also be set to the correct gap setting for the engine. Most engines will have use a magneto to provide the spark for the explosion. These will have the points built directly into the magneto, as well as the condenser, distributor, and coil. Low tension magnetos are also available and are less prone to problems, but are more expensive. Solid state ignitions are also available. The starter is provided with electricity either by the electrical grid or a battery. Batteries should also be cared for so that they last as long as possible. That includes protecting them from weather and extremes of temperature. Other parts of the electrical system for an engine can include an alternator or generator, distributor, voltage regulator, and whatever wiring is needed. Oil and Lubricating System Oil is used to prevent friction throughout the engine, and it’s circulated throughout the engine outside of the cylinder. The amount of oil and the pressure at which it’s run through the system are two critical aspects of the lubricating system. The oil amount is measured by the level of oil in the engine. A safety measure is included in the form of a float and switch. If the oil level drops too low, the engine will shut down. A similar, but more complex, safety system shuts the engine down if the oil pressure falls too low. The correct weight of oil must also be used for each engine. Weight in this case refers to the oil’s viscosity, or how easily it flows. 10 weight oil flows much more quickly than 40 weight oil, and oil will flow at different weights depending on its temperature, so it will often be labeled with 2 numbers. Oil that is 10-40 will flow as fast as a 10 weight oil when it’s cold, but will travel as slowly as a 40 weight oil when it’s hot. Oil will also have a number of different additives in it, so the oil is doing more than providing lubrication. Some additives will improve the viscosity so that the oil sticks to the engine parts and does a better job of preventing friction. Other additives prevent rust or help to clean the engine. Oil and the additives it can contain will breakdown over time, especially when exposed to heat. The oil and the oil filter will need to be replaced regularly. When doing so, you should use the correct weight and type of oil for each different engine. Antifreeze and Coolant System The coolant system removes heat from the engine by circulating water and antifreeze. The coolant then is run through a radiator that is open to passing air. The air draws heat from the coolant in the radiator, so that it can be cycled back into the engine to be used again. The temperature of the coolant is monitored, and if it gets too hot the engine is automatically shut down. As the coolant used in most engines is water, it is in danger of freezing in cold weather. To prevent that, antifreeze is added to the water. The main component in almost all antifreeze is ethylene glycol. It’s also effective as a rust inhibitor and as a lubricant for the coolant system. Antifreeze also has a higher boiling point, at 263 degrees Fahrenheit, so that the coolant transfers heat more efficiently. While it’s unlikely for the coolant to freeze during a hot summer, it’s wise to continue to add antifreeze for other benefits it has. The standard ratio of water to antifreeze is 1/3, or one part antifreeze to two parts water. That mixture is optimized for freezing temperatures to 0 degrees Fahrenheit. Different mixtures are better for other climates. The table below offers some mixtures optimized for a range of temperature ranges. |Ratio of Antifreeze||Ratio of Water||% of Antifreeze||Freezing point The fuel that an engine uses will determine a lot about how it functions. An engine powered by gasoline will need a carburetor or fuel injector. Engines using other possible fuels like butane or natural gas will also need carburetors. Diesel engines will need a fuel injection system. You’ll have to maintain engines powered by many different fuels, so it’s a good idea to be familiar with as many as possible. Just like your car, the gas you put in the engines on your lease should be of the right type and grade. Using the wrong grade of fuel can lead to reduced efficiency, as well as some longer term problems. Gas can come with a number of different additives that do many of the same things as those added to oil. Some will prevent acids from forming through condensation. Carburetors can become clogged over time, so some additives are designed to help clean them. Water in the fuel line can lead to rust or freezing damage, so some additives help water pass through the system harmlessly. Some will also improve the performance of the engine. Octane improvers will slow down the burning of the fuel-air mixture in the cylinder. That allows the power generated to be used more effectively, and also reduces the wear on the engine itself. Higher octane fuels can lead to better performance and a cleaner engine, but are also usually more expensive. Diesel powered engines don’t use spark plugs, as the heat generated by the piston compressing the air is enough to ignite the fuel. Fuel is injected at the moment the piston reaches the top of the cylinder and burns instantly. Because of the heat that is generated, diesel fuel has oil mixed into it as lubricant. Steel expands when hot, so diesel engines aren’t built with tolerances as close as gasoline engines. They’ll make more noise when they’re first started up and grow quieter as the engine heats up. Diesel engines have glow plugs that heat up the air in the cylinder so that the engine will start. Gas is an obviously appealing power source, particularly when it’s a product of the well you’re pumping. Gas from the well will often contain a lot of water. This has to be removed from the gas before it can be used as fuel. When you’re using the gas directly from the well, you’ll need to pass it through a regulator, so that it’s at the correct pressure for your system. The gas can then be passed through a wash tank to remove water. Water in the fuel system can freeze in cold weather. It will condense out of the gas and the collect in low points in the system, eventually blocking the line. Adding a small, custom modified tank to the system to collect the moisture can help prevent blockages, as long as the water is blown out of the tank regularly. Figure 3. The layout of a natural gas fuel system, showing the flow from well to engine including all major components. (courtesy Arrow Specialties, Co., Inc.) In Figure 3, you can see an example layout of the system. A scrubber and tank may be needed when supplying an engine with gas. The scrubber will remove water, though if the gas supply is dry enough to be used as fuel you don’t need to add a scrubber to the system. The volume tank will keep the fuel supply steady and the pressure steady; 6 or 7 ounces of pressure is usually required. Larger engines will have a greater fuel need, and so a second volume tank may be needed. The rise, the pipe projecting upward in Figure 3 between scrubber and volume tank, is used to prevent a sudden rise in pressure or amount of gas. Is your appetite for oil & gas operating knowledge insatiable like ours? 😀 If so, check out these related articles, Lease Electrical System and Motors In Oil & Gas Production and, Utility Pumps For Oil & Gas Production – they’ll be sure to pump you up!!!
Plant4BorneoElephants: ‘Whilst sitting in a Tawau plantation’s countryside lodge’ A word of explanation for this title: In August 2018, I was staying in the spacious visitors’ accommodation of a Tawau plantation whilst on a trip organised by IStopBorneoWildlife, a Sabah-based NGO. I was there as part of a conservation project developed between the plantation and the local NGO that has two aims: to promote wild elephant spotting for tourism and to create a wildlife corridor of forest trees to connect two neighbouring wildlife reserves. In early 2018, 1StopBorneoWildlife established that a herd of approximately 80 wild elephants was roaming within the plantation of softwood trees, for example Acacia, and the local management there were taking the enlightened approach of seeing the elephants as a potential asset rather than a threat to their staff and trees. In addition, there already existed, in outline at least, a 13.89-kilometre wildlife corridor within the plantation that was intended to link the two nearby forest reserves, Ulu Kalumpang and Ulu Segama. And so the partnership was born: Plant4BorneoElephants. Now the plantation provides overnight accommodation, transport for observing the elephants and the knowledgeable staff to track them. 1StopBorneoWildlife provides the visitors in conjunction with a local tour operator, Adventure Alternative Borneo, but also ensures close liaison with the relevant plantation staff. However, there is another component to this conservation partnership: the development of the wildlife corridor in which the saplings of local native trees are cultivated using the plantation’s expertise whilst the visitors travel to the corridor and plant them. A typical visit, therefore, involves open-topped vehicle rides along the plantation roads to track and spot the elephants. Sightings cannot be guaranteed but local knowledge and a coordinated network of ‘spotters’ from among the plantation staff, providing ‘real time’ information to the driver, considerably increases the likelihood of seeing these surprisingly elusive animals. After an evening meal, a night’s sleep and breakfast, visitors should now be ready to select tree seedlings from a large range of species cultivated in the nearby nursery. These may be wild fruit such as durian or figs, or perhaps one of a number of dipterocarp, for example Shorea species, or non-dipterocarp species. Armed with spades and under the direction of plantation staff, the visitors are driven to a designated area in the corridor where positions for planting have already been mapped out. The input of local people within the plantation should not be underestimated in ensuring proper and planned planting is carried out. By actually planting tree seedlings, visitors certainly gain insight into the challenges of practical conservation! The two components of the project both address issues that involve the interaction between elephants and humans. The elephant in question is Elephas maximus borneensis, the Borneo Pygmy Elephant of which approximately only 1,500 remain in the wild, mostly in Sabah alone. In other words, it is a critically endangered species and is a genetically distinct subspecies of the Asian elephant which in total has a maximum wild population of about 50,000 surviving in Asia. The pygmy elephant is a species of Bornean lowlands and valleys which is, sadly, the habitat typically exploited by humans for logging, palm oil cultivation and other forestry activities such as softwood cultivation. It is a herbivore, feeding on fruits, grasses and some trees such as palms. A herd may need an area of lowland forest of perhaps 300 square kilometres for feeding and to maintain breeding relationships. These features of its biology mean that conflict with the expanding human population in Borneo is inevitable. It is now confined to the NE tip of Kalimantan and southern and eastern parts of Sabah; it has been eliminated completely from Sarawak. The elephant/human conflict has become sharper as vast areas of rainforest have been claimed for palm oil cultivation in the last 30 years. This expansion of palm oil cultivation not only reduced the elephants’ natural habitat but has also had the effect of isolating areas of surviving primary forest; both of these features of forest destruction were bound to increase contact and conflict with humans. For more information see the following website: A visitor to the plantation will immediately be struck by the large acreage of softwood trees with their uniformity and lack of diversity of plants and animals they support. The plantations are grown to produce the maximum yield of wood as rapidly as possible before felling and so a monoculture is the result. Although the trees are fast growing and prevent the growth of all but a few plant species in their shade, it does seem that elephants are able to feed on fast growing plants such as grasses within and on the borders of these plantations. There may also be other species present after recent reported sightings of 3 hornbill species, notably the rhinoceros hornbill, and other mammals such as red leaf monkeys, sambar deer and leopard cats. In contrast, the creation of the 13.89- kilometre wildlife corridor, with a total area of 1,067 hectares, demands a very different planting strategy so as to develop the diversity that characterises the primary forest protected in the two reserves being linked. There is no quick fix for this, long term management is required. A whole range of tree seedlings must be planted and, since these are slower growing, they must be protected for decades as they and the ecosystem they support become established. The history of the recent decline of so many animal species in Borneo clearly demonstrates the need for solutions to prevent their extinctions. The need to educate people, not just in the palm oil producing countries of Indonesia and Malaysia but throughout the world, about rainforests, their biodiversity and destruction is paramount. One way of doing this is through conservation tourism that has the additional, potential benefit of raising money for further conservation. The role of local tour operators and conservation NGOs such as those involved here, is important for tourism but they may also be vital in the education of people of all ages and backgrounds in the issues of conservation. It is pleasing to hear that Datuk Christina Liew, the State Minister of Tourism, Culture and Environment praising IStopBorneo Wildlife for initiating this elephant conservation project that demonstrates how Sabah’s elephants can be an asset for the State and not a liability as so often reported. As reported in The Star Online (November 24th 2018) she said ’this NGO, founded in 2016, came up with a great idea of working with oil palm plantations for its initiative known as Plant4BorneoElephants.’ For more information on 1StopBorneoWildlife go to: https://www.1stopborneo.org/ Those involved in agricultural businesses, such as the company involved here, must also acknowledge wider responsibilities beyond being profitable. To be fair, for example with this project, this plantation, at least, seems to be prepared to do just that. The company’s timber concession belt in Tawau is part of a more substantial area of timber and palm oil cultivation totalling upwards of 40,000 hectares and so this project is being developed in just a small part of this. Long may it continue its support. The future: educational and tourism potential Such an initiative should benefit all parties: as Datuk Christina Liew has noted: ‘ 1StopBorneo Wildlife’s undertaking has boosted the local Tawau economy by encouraging tour operators to take tourists to the plantation for elephant viewing and planting seedlings.’ At the same time, the need for new jobs such as drivers and community guides has boosted job opportunities for local people. This partnership can be sustainable and relevant: the elephants need to move out from the plantation and the development of the wildlife corridor will help to achieve this. Visitors are active contributors to the partnership when they plant seedlings; they gain an insight into the challenges of conservation, rather than merely being passive observers. However, the enormous educational potential of such a partnership should not be overlooked: both school and university students will benefit from visits, both providing the energy for tree planting whilst also observing elephants in the wild. It offers a model for other plantations in Sabah to explain their approaches to balancing conservation with the core business of large scale cultivation of commercial crops. It is an important opportunity for plantation owners to contribute to conservation. The vulnerability of mega fauna such as the Borneo pygmy elephant to the consequences of human contact is clearly illustrated by the reports of so many elephant deaths in Sabah this year. This new conservation project in this small part of Sabah offers a way forward and coincides with the more positive news, that in this changing political era, local bodies may finally be prioritising elephant conservation. Lee, S. (2018) Jumbo boost for Sabah tourism The Star Online November 24th 2018
१६ सितम्बर २०१९ | डॉ पूर्णिमा शर्मा (इस आर्टिक्ल को ४७७ बार पढ़ा जा चुका है) पितृपक्ष चल रहा है | सभी हिन्दू धर्मावलम्बी अपने दिवंगत पूर्वजों के प्रति श्रद्धा सुमन समर्पित कर रहे हैं | हमने भी प्रतिपदा को माँ का श्राद्ध किया और अब दशमी को पिताजी का करेंगे | कुछ पंक्तियाँ इस अवसर पर अनायास ही प्रस्फुटित हो गईं... सुधी पाठकों के लिए समर्पित हैं... द्विविधाओं की लहरों में डूबता उतराता सा... घिरा हुआ कुछ कुछ जाने कुछ अजाने / कुछ चाहे कुछ अनचाहे / नातों से... सोचने लगता है / नहीं है कोई उससे अधिक धनवान इस संसार में... जब सहलाते हैं कितने ही स्पर्श प्रेम से / अनुराग से... एक अपरिचित सी आशंका / कहीं ये सब स्वप्नमात्र तो नहीं... लेकिन साथ ही होती है एक अजीब सी सिहरन भी / अविश्वास की... मुखौटे चढ़ाए हुए उन "अपनों" की प्रकृति का... तभी तो बहाता है अश्रु / घिरा हुआ अपनों की भीड़ में भी... हो जाता है उदास / "अपनों" के स्नेहिल स्पर्शों से भी... ममतामयी गोद जनयित्री की... विश्वास से परिपूर्ण नेत्रों का जनक के भी... नहीं हैं जो आज शरीर रूप में निकट मेरे... और करते हैं मान मेरे हर प्रयास पर... और प्राप्त कर सकूँ लक्ष्य को अपने / बिना किसी बाधा के...
THE STEAMGARD SYSTEM® applies the principles of two-phase flow technology to a custom modified Venturi Nozzle configuration Steam and condensate pass through the stainless steel STEAMGARD Condensate Removal Device by a series of staged steps. First, the denser liquid (condensate) throttles the specialized Venturi Nozzle, which keeps the steam from escaping. Erosive eddy currents are eliminated and any contamination is carried away with the condensate. Because condensate is created continuously in a working steam system, the condensate continuously chokes the Nozzle from losing steam as the condensate is effectively removed. THE STEAMGARD SYSTEM® has no moving parts. Its components include: - A custom modified Venturi Nozzle to remove condensate - A high quality strainer and secondary screen to protect the stainless steel nozzle - A blowdown valve to clean the strainer of debris The Science of the Nozzle STEAMGARD Condensate Removal Devices utilizes a custom modified Venturi Nozzle to separate the steam from the condensate. In any piece of steam equipment before the steam trap, there is always a two-phase mixture (i.e., steam plus condensate). - At 1 Barg (14.5 psig), condensate is over 850 times denser than steam (density of condensate being 943 Kg/m3 versus 1.1 Kg/m3 for steam). - As the condensate density is much higher than steam, it moves at a much slower velocity through the venturi nozzle as compared to steam. - From the basic physics of two-phase flow, we know that the heavier phase (in this case condensate) preferentially passes through the nozzle as compared to the lower density steam. - In a system with both steam and condensate present, the slow moving condensate holds back the live steam and condensate is discharged preferentially through the nozzle. - Steamgard units always have a condensate seal at the inlet of the venturi nozzle. - For variable flow applications, the regulation of the capacity in a Steamgard venturi nozzle is achieved by the generation of flash steam as the hot condensate passes through the nozzle. - Flash steam is formed in the venturi due to the pressure drop between the inlet and the outlet of the unit. - This is per Bernoulli’s law according to which the fluid flow velocity (i.e., kinetic energy) through a pipe restriction must increase. The increased kinetic energy comes at the expense of fluid pressure (i.e., potential energy). - The pressure drop occurs as a result of the condensate return system always being at a lower pressure than the steam. - As the saturated condensate travels from the inlet to the outlet of the nozzle, flash steam is produced which blocks any flow of live steam. - 3,000 PSIG/ 206 BAR (g) - and 950 °F/ 510 °C - from 0.5 inches to 4.0 inches - Long Life with Simple Maintenance - Each custom modified Venturi Nozzle is manufactured using Electrical Discharge Machining (EDM) and is ISO 9001 certified for quality control and manufacturing. - Precise design and manufacturing of Nozzle staging, lengths and angles. - Designed and manufactured with a low RA factor for the smoothest possible surface. - Only the highest grade stainless steel is used in our Venturi Nozzles ensuring there is never a degradation of the nozzle.
Agricultural Registers and their meanings are listed and explicated in this article. We hope you find this helpful and informative. Read Also: Forms of Agriculture Advanced Agriculture: a style of agriculture that brings in science and technology for the production, processing, storing, and marketing of agricultural products. Anther: this is a part of the flower that stores the needed pollen for fertilization Artificial: Anything on the farm that is made by man. Barn: This is where tuber crops are stored on the farm Carbohydrates: A source of food and energy for animals and man which contains a lot of sugar and starch. Read Also: Agriculture and Nigeria Economic Development Cereal: These are grains of grasses such as wheat, rice, or oats. Compound leaf; a leaf with more than one leaflet on the same leaf stalk. Crop: a term used in place of plants. Cultivation: This is the process of making the soil ready for the planting of crops. Domesticated: a term used for animals that have been brought under control and are not wild. Economic empowerment; is a situation where agriculture is used in order to provide money to buy other necessities of life. Export: This is the act of selling out agricultural products to another country. Fertile: this is a situation when the soil has the means or ability to produce crops when planted on it; fertile soil is naturally rich with all the materials that the soil may need for adequate germination. Fertilization: This is the process of making something very fertile for example, adding chemicals or organic fertilizer to the soil to improve its quality. Forage crops: These are crops that are grown to be used for food for cattle and other livestock. Germinate: This means to grow or cause something to grow. Grading: This is the process of selecting the produce of agriculture based on their sizes, shapes, purity, and other factors in order to ensure high-quality standards. Harvest: this is the gathering of crops that are fully matured and ripened. Higher Plants: These are plants that are multicellular and are highly developed and also possess a complex structure. Read Also: Importance of Agriculture in Nigeria Humidity: This is the amount of water in the air at a particular point in time. Import: Bringing in goods from outside the country. Insecticides: these are chemicals that are used by the farmer to destroy insects that are pests. Irrigation: this is a method of bringing in water through artificial canals and ditches to the areas on the farmland which do not have water. Legumes: this is a plant that has so many pods (for example, peas, and beans). Livestock: this often refers to animals that are kept and reared for agricultural purposes. Mineral Salt: These are particles in the soil that are needed for the good growth of plants. Node: This is the place on the stem of a plant from which the leaves grow. Organic: This is formed from living things. Organic fertilizers are usually formed from the waste products of animals. Photosynthesis: This is the process by which green plants use sunlight to convert carbon dioxide which is taken from the air and water into plant food. Processing: these are the changes that are made to agricultural products before they are finally sold or consumed. Propagate: to reproduce a line of animals or plants through breeding and reproduction. Stigma: This is the part of the flower that revived the pollen in the process of pollination. Read Also: Step To Step Guide To Start Beans Farming Stomata: These are the pores in the outer layer of cells of a leaf that controls the passage of gasses in and out of the plant. Tillers: these are shoots that arise from the base of a stem. Tubers: these are fleshy underground stems or roots e.g. Potatoes, cassava, yam, etc.
Artificially printed tissue may one day revolutionize how diseased organs are treated, fractured bones are fixed, and drugs are tested without human subjects. A great deal of progress has already been made toward this goal, but a major hurdle, that of integrating vasculature and bringing together different cell types into a functional whole, has significantly limited that progress. Now researchers at Harvard University report in journal Advanced Materials on a new method of 3D printing constructs made of three different cell types, including ones that line blood vessel walls. Because any substantially large chunk of tissue requires oxygen to penetrate into its interior, the vessels within the construct allow for much larger pieces of printed tissue to be created. Some details from a Harvard news release: To print 3D tissue constructs with a predefined pattern, the researchers needed functional inks with useful biological properties, so they developed several “bio-inks”—tissue-friendly inks containing key ingredients of living tissues. One ink contained extracellular matrix, the biological material that knits cells into tissues. A second ink contained both extracellular matrix and living cells. To create blood vessels, they developed a third ink with an unusual property: it melts as it cools, rather than as it warms. This allowed the scientists to first print an interconnected network of filaments, then melt them by chilling the material and suction the liquid out to create a network of hollow tubes, or vessels. The Harvard team then road-tested the method to assess its power and versatility. They printed 3D tissue constructs with a variety of architectures, culminating in an intricately patterned construct containing blood vessels and three different types of cells—a structure approaching the complexity of solid tissues. Moreover, when they injected human endothelial cells into the vascular network, those cells regrew the blood-vessel lining. Article in Advanced Materials: 3D Bioprinting of Vascularized, Heterogeneous Cell-Laden Tissue Constructs…
Responding to the COVID-19 pandemic required extreme measures by ORs throughout the nation to address the crisis at hand, including a recommendation from CMS, the U.S. Surgeon General, and many other renowned medical groups to temporarily suspend and cancel elective surgeries and procedures. Many healthcare organizations and hospitals responded appropriately, postponing or cancelling nonessential surgeries. Fortunately, we’ve started seeing signs that the first wave of this pandemic might be coming to an end, and organizations are starting to create project plans for resuming elective surgeries and procedures. The American Society of Anesthesiologists (ASA), the American College of Surgeons (ACS), the Association of periOperative Registered Nurses (AORN), and the American Hospital Association (AHA) have released a joint roadmap and list of considerations for resuming elective procedures after the COVID-19 pandemic. Nearly 40% of the patients were unwilling to schedule their procedure at this time. The time is now to begin planning for how ORs should address the overflowing case demand and backlog of surgical procedures that were cancelled due to COVID-19, as well as the normal surgical demand. We’ve outlined seven considerations for restarting your ORs and addressing elective case backlog. It’s important to note that your backlog is a combination of cases that have been cancelled as result of COVID-19 as well as cases that were in the pipeline at surgeon offices that were never put on the schedule due to the pandemic. 1. Put patient and staff safety first. The number-one consideration to factor into resuming elective surgeries is patient and staff safety. Hospitals should aim to book surgeries at least 14 days out, 10 at a minimum. Patients being added to the surgical schedule should automatically be tested for COVID-19. If their results come back positive, they should be removed from the schedule and reschedule only once they’ve tested negative, twice. Once a patient’s surgery has been scheduled and they’ve been tested for COVID-19, the patient should then be instructed to self-isolate from then until their procedure date. 2. Create a COVID-19-specific cancellation code. It’s important to understand the number of cases that have been or will be cancelled due to COVID-19, for both reporting purposes and beginning to address surgical case backlog. Once you’ve created a COVID-19-specific cancellation code, take a retrospective look at all of your cases since the pandemic started (we’d recommend at least March 15) and re-code cancellations that are particular to COVID-19. This can help identify cases that still need to be done and will allow your surgical team to appropriately estimate OR demand in the coming months. 3. Evaluate your OR demand and capacity. Create a pipeline report by surgeon and specialty that identifies projected cases and case time needed to complete backlog cases. Backlog cases are those that were canceled by the hospital due to COVID-19 and those that were never scheduled by the surgeon’s office. Evaluate the case times, turnover times and needed hours of block to complete the backlog for each surgeon. This will be important as your OR adjusts blocks and times during the restart period. Additionally, evaluate your OR capacity during current primetime hours and consider extended hours with a goal of optimal utilization to address the current backlog of cases. 4. Prioritize cases using a scoring system, such as the Medically Necessary Time-Sensitive (MeNTS) scale. A scoring system will help with determining how to prioritize cases and when to proceed with certain surgical operations in light of resource limitations and exposure risks posed by COVID-19. The MeNTS prioritization scale’s methodology, developed by investigators and surgeons at the University of Chicago, addresses elective surgical procedures and guides both surgeons and OR leaders across different specialties in prioritizing case types and patient safety. There are 21 factors considered in the scoring system, scored on a scale of 1 to 5. Thus, a total score can range from 21-105. The higher the score, the greater the risk to the patient, the higher the utilization of healthcare resources, and the higher the chance of viral exposure to the healthcare team. A full list of factors and a sample MeNTS worksheet can be found here. This system is ideal for any OR, not just academic medical centers or large health systems. The assessment of resources and risk is applicable anywhere, which makes this a strong tool for prioritizing and determining cases to go forward with amid the COVID-19 pandemic. 5. Develop individualized plans to adjust blocks based on prior utilization and extend release times to accommodate a backlog of deferred cases. There needs to be a concentrated effort to enforce block policies to ensure optimal efficiency and max utilization to accommodate a backlog of cases. OR leaders should consider extending block release times out by at least seven days (with cardiac and trauma being the exceptions to this) to allow specialty teams more flexibility in scheduling as the surgeon offices begin booking cases again. Take some time to review your block utilization numbers from January and February. OR leadership should consider temporarily releasing blocks that had less than 50% utilization in January and February. Additionally, consider a “Surgery Saturday.” This would require opening a couple of your ORs on a Saturday for high-volume surgeons with consistent case-time procedures. This would be for shorter cases like hernia repairs or similar cases (less than 45-minute case times). 6. Consider a modified visitor policy specifically for surgery patients. A local hospital began contacting patients to schedule their canceled elective surgery and found that nearly 40% of the patients were unwilling to schedule their procedure at this time. Cases that would require an overnight stay were even more likely to be declined with the visitor restriction policy being cited as one of the driving factors. Fear and safety concerns are understandable and should be addressed with good policy and precautions. If you’re leading an OR at a hospital that doesn’t have any active COVID-19 cases or that can restrict cases to a specific floor or floors, it might be worth considering a modified policy for visitors. Testing and self-quarantine requirements used to keep surgical patients safe could be extended to a specific visitor for the purposes of a planned elective case. 7. Use CMS’ “Hospital without Walls” exemption to complete cases at an ambulatory surgery center (ASC) instead of the hospital. Maximize volume of surgical cases in a way that promotes patient safety and preserves resources at the main hospital. This will also protect inpatient capacity at your main campus. The Hospital without Walls initiative allows currently enrolled ASCs to temporarily enroll as hospitals and provide hospital services. During this time, ASCs can bill procedures as if they’re being performed at the hospital. For more information, ASCs that want to enroll to receive temporary billing privileges as a hospital should call the COVID-19 Provider Enrollment Hotline or review the CMS Medicare provider enrollment relief FAQ. If you’re interested in learning more and collaborating with other OR directors facing the same challenges, consider joining one of our OR restart collaboratives. We’re hosting a collaborative session for hospital leaders, OR directors, and surgeons to discuss current issues and how to restart ORs amid the continuing COVID-19 pandemic. If you’re interested in participating or would like to learn more, please contact [email protected] or [email protected]. For more information on this topic, check out our article at Fierce Healthcare.
Rebalance vs. HODL: A Technical Analysis The intention of this study is to paint a fair picture of how rebalancing as a strategy stacks up to HODLing. In order for this to happen, we thought carefully about how we would design the backtests, data, and variables. Trades & Data A complete year of market data was collected from exchanges. This data was used to evaluate the cost of each trade at the exact time a rebalance would have been performed. Additionally, all trades included a .25% fee which is standard for most exchanges. So, a trade from LTC to SNT would trade from LTC to BTC and then BTC to SNT. In this instance, both trades incurred a .25% fee. This allows us to create the most accurate model possible for rebalancing performance. The first variable that is necessary for this study is the rebalance period. A rebalance period is the specific amount of time between each rebalance. So, a period of 1 day results in a rebalance every single day at the exact same time. The purpose of varying this value is to determine if the frequency of rebalances affects the performance of a portfolio. In this study, we selected rebalance periods of 1 hour, 1day, 1 week, and 1 month. Learn more about rebalancing for cryptocurrency. The second variable we decided to investigate for this study was the number of assets in a portfolio. The hypothesis was that the number of assets in a portfolio has a strong influence on the performance. This hypothesis was tested with 5 groups of asset sizes. Since 2 is the smallest number of assets that will produce any difference when comparing rebalancing and HODLing, we started with a 2 asset portfolio. Then, we increased by 2 to obtain 2, 4, 6, 8, and 10 as the number of assets in each portfolio group. Learn more about how the number of assets in a portfolio affects performance. In order to determine which assets would be considered during the process of constructing a portfolio, we used a cross section of Bittrex and Poloniex. This means we took all of the assets from Poloniex for which we had 1 year of data and compared them to the list of Bittrex assets for which we had a year of data. Any asset which was in both lists was included in our pool for the selection process. When a portfolio was constructed, assets were randomly selected from the pool to create a portfolio. While our study randomly selects assets, we strongly discourage this as a strategy for creating a portfolio. Learn more about how to successfully build a strong portfolio. A backtest is the process of using the trade data from the exchange to simulate how a strategy would have performed over a given amount of time. This is often used to test the viability of a strategy by running it through these large data sets. In this study, we used backtests to compare the results of rebalancing to those of HODL. The number of backtests we ran for each portfolio size and rebalance period pair was set to 1000. This was determined to be sufficiently large to produce an obvious trend. Read more about backtests or run your own. Now that we know how the study was set up, let’s walk through the entire process. First, the rebalance period was set to 1 hour and the number of assets was set to 2. This means the portfolio would contain 2 assets and rebalance 1 time every hour. Next, 2 assets were selected at random from the pool of assets. If there were no duplicates, the backtest was run. Once complete, the software then randomly selected 2 new assets at random and ran another backtest. This process continued until it successful ran 1,000 backtests. Once complete, the number of assets was increased from 2 to 4 and 1,000 more backtests were run. This process continued until each combination of number of assets and rebalance periods were backtested. 2 Asset Portfolio A two asset portfolio represents the most simple option for a portfolio. In this instance, the cryptocurrencies simply trade back and forth to each other during each rebalance period. We can see from these histograms that the shorter rebalancing periods result in a larger spread in performance. There are significantly less outliers for shorter rebalancing periods and the results are consistently higher. As the rebalance period increases, the spread actually decreases. This results in less variance in results, but a higher observance of outliers. This suggests higher periods produce lower returns consistently, but also produce more sporadic outliers. The portfolios which used a 1 hour rebalance period outperformed buy and hold by the largest difference of 93%. 4 Asset Portfolio Continuing the trends from the 2 asset portfolio study, we see that shorter rebalance periods have larger spreads in performance in the 4 asset portfolios as well. This results in fewer outliers and a significantly higher median performance than the longer rebalance periods. It can also be observed that the highest performing portfolios all utilized a 1 hour rebalance period. This is even the case when including all outliers. A period of one hour performed the best with a 177% gain OVER buy and hold. 6 Asset Portfolio We observe from the 6 asset portfolio results that the trends discussed in the 2 and 4 asset portfolios continue. This includes the larger spread for shorter rebalance periods and a higher average performance for shorter rebalance periods. At this time, we can also begin to conclude that there is an increasing spread between the 1 hour rebalance period and the 1 month rebalance period as we increase the number of assets in the portfolio. We can keep that in mind as we continue the study. A portfolio which contains 6 assets and has a rebalance period of 1 hour outperformed HODL by 203%. 8 Asset Portfolio We observe from the 8 asset portfolio results that the trends discussed in the 2, 4, and 6 asset portfolios continue. This includes the larger spread for shorter rebalance periods and a higher average performance for shorter rebalance periods. What we can also see is that there is only one histogram in this study of 8 asset portfolios that contained results which performed worse than HODL. This can be seen in the bottom right chart which represents the portfolios which used a 1 month rebalance period. The median 8 asset portfolio which rebalanced every 1 hour outperformed HODL by 224%. 10 Asset Portfolio We observe from the 10 asset portfolio results that the trends discussed in the 2, 4, 6, and 8 asset portfolios continue. This includes the larger spread for shorter rebalance periods and a higher average performance for shorter rebalance periods. We can also see from these results that only 10 portfolios out of 4,000 performed worse than HODL if they had rebalanced even 1 time each month. This means if you randomly selected 10 assets and rebalanced at least once a month, you would have had a 99.75% chance of outperforming buy and hold over the last year. This is truly incredible. The median performance for a portfolio with 10 assets and a rebalance period of 1 hour was 234% BETTER than HODL. Now that we have all of the data, we can simplify the results into a 4 x 5 grid that illustrates the performance of each portfolio and rebalance period. Since the upper bound on most graphs is much higher than the lower bound, we will calculate the median. This also means that 50% of the portfolios were above and 50% of the portfolios were below this value. So, if when creating your portfolio, you randomly selected assets without performing any research, you would have a 50% chance of performing better than the listed value. Also, the listed value is the percent gain over buy and hold. So, a value of 10% would mean rebalancing performed 10% BETTER than HODL. We can draw two major conclusions from this grid. First, we have obvious correlations between the rebalance period and the performance. As the rebalance period becomes shorter, the performance of the portfolio increases. A second correlation we can see is between the number of assets and performance. As the number of assets in the portfolio increases, there is an increase in performance. Therefore, the best performing portfolios were those that have both a short rebalance period and a large number of assets. To round out the complete comparison, we will combine every backtest to create an overall comparison. The results show a median performance increase of 64% over all portfolio sizes, rebalance periods, and coin selections. Tax Implications (US Specific) According to the latest news, crypto trades are taxed as short-term capital gains at the rate of your current income bracket if the assets were held for less than a year. Long-term capital gains will be taxed at a discount when assets are held for more than a year. Since there are a lot of misunderstandings revolving around taxes, I will try to break down some of the implications here. All calculations will be based on an individual income of $120,000. An individual making $120,000 is well within the top 10% of incomes in the US. They are also in a federal income tax bracket of 24%. This means any short-term capital gains will be taxed at 24%, which is equal to the personal income tax. That same individual making $120,000 will pay long-term capital gains at 15%. We can see quickly that there is a 9% difference in taxes between long and short term capital gains. We can compare this difference to the 64% boost in returns observed through rebalancing. What we see is that rebalancing significantly outperforms HODL even after factoring in tax implications of frequent trading. In fact, 92% of all portfolios which rebalanced over the past year beat HODL, after taxes. That’s not the entire story, however. Rebalancing only trades a portion of the portfolio at any given time. This means part of a portfolio which uses rebalancing would not have been traded by the end of one year. These untouched portions can be taxed as long-term capital gains, reducing the overall taxes that are incurred as a result of rebalancing. The amount can be quantified by examining the volatility difference between all cryptocurrencies over the last several years. This would give us an idea of what percentage of a portfolio would typically be considered long-term capital gains. Since a proper simulation would require careful design, we will save this analysis for another post. There are two major relations we can draw from this study. The first relation is that increasing the number of assets increased the performance of a portfolio.The second relation is that decreasing the rebalance period (increasing rebalance frequency) increased the performance of a portfolio. Therefore, the ideal portfolio was rebalanced frequently and also contain numerous assets. It should be remembered that all of these portfolios were selected on a completely random basis. There was no research or elimination process when determining which assets should be incorporated into the portfolios. There is a significant amount of improvement potential for an individual who actively researches and selects promising assets. Rebalancing beat HODL by a median of 64%. After taxes, this represented 92% of all possible cryptocurrency portfolios. Rebalancing with Shrimpy Now that we have determined that rebalancing was objectively better than HODLing, we need a way to capitalize on this knowledge. This is where the Shrimpy application can help. Shrimpy is a completely free service that automatically rebalances your portfolio. However, that’s not all it does. It is the easiest way to manage your portfolio. Quickly select assets, instantly allocate a portfolio, and monitor your investment over time. It’s the less stress, more gains solution to portfolio management. Sign up today by clicking here. If you still aren’t sure, try out the demo to see everything we have to offer! _ _ _ Leave a comment to let us know your experiences with rebalancing! The Shrimpy Team
न्याय की कठिन राह -राजेन्द्र बंधु- -विशेष आलेख: ४२ वर्षोंं तक कोमा मेंं रहने वाली दुष्कर्म पीड़िता की मृत्यु के बाद उठे सवालों पर केन्द्रित आलेख नृशंस यौन हमले की पीड़िता अरूणा शानबाग अब जिन्दगी के आजीवन कारावास से मुक्त हो चुकी है, जबकि अपराधी सिर्फ सात साल की सजा भुगतकर इस समाज का हिस्सा बना हुआ है। नईदिल्ली रेड मोर
############## # Add Binary # ############## ''' Given two binary strings, return their sum (also a binary string). For example, a = "11" b = "1" Return "100". ''' class Solution: def addBinary(self, a, b): """ :type a: str :type b: str :rtype: str """ return bin(int(a,2)+int(b,2))[2:] ################### # Add Two Numbers # ################### ''' You are given two non-empty linked lists representing two non-negative integers. The digits are stored in reverse order and each of their nodes contain a single digit. Add the two numbers and return it as a linked list. You may assume the two numbers do not contain any leading zero, except the number 0 itself. Example Input: (2 -> 4 -> 3) + (5 -> 6 -> 4) Output: 7 -> 0 -> 8 Explanation: 342 + 465 = 807. ''' # Definition for singly-linked list. # class ListNode: # def __init__(self, x): # self.val = x # self.next = None class Solution: def addTwoNumbers(self, l1, l2): """ :type l1: ListNode :type l2: ListNode :rtype: ListNode """ resoult=res=ListNode(0) carry=0 while l1 or l2 or carry: if l1: carry+=l1.val l1=l1.next if l2: carry+=l2.val l2=l2.next div,rem=divmod(carry,10) carry=div res.next=ListNode(rem) res=res.next return resoult.next ############### # Add Strings # ############### ''' Given two non-negative integers num1 and num2 represented as string, return the sum of num1 and num2. Note: The length of both num1 and num2 is < 5100. Both num1 and num2 contains only digits 0-9. Both num1 and num2 does not contain any leading zero. You must not use any built-in BigInteger library or convert the inputs to integer directly. ''' class Solution: def addStrings(self, num1, num2): """ :type num1: str :type num2: str :rtype: str """ carry=0 res=[] num1=list(num1) num2=list(num2) while len(num1)>0 or len(num2)>0: if len(num1)>0: carry+=ord(num1.pop())-ord('0') if len(num2)>0: carry+=ord(num2.pop())-ord('0') div,rem=divmod(carry,10) carry=div res.append(rem) if carry: res.append(carry) return ''.join([str(i) for i in res])[::-1] #################### # Multiply Strings # #################### ''' Given two non-negative integers num1 and num2 represented as strings, return the product of num1 and num2. Note: The length of both num1 and num2 is < 110. Both num1 and num2 contains only digits 0-9. Both num1 and num2 does not contain any leading zero. You must not use any built-in BigInteger library or convert the inputs to integer directly. ''' class Solution: def multiply(self, num1, num2): """ :type num1: str :type num2: str :rtype: str """ l1=len(num1) l2=len(num2) if l1<=l2: rnum1=num1[::-1] rnum2=num2[::-1] else: rnum2=num1[::-1] rnum1=num2[::-1] a=0 for idx1 in range(len(rnum1)): n1=ord(rnum1[idx1])-ord('0') carry=0 for idx2 in range(len(rnum2)): n2=ord(rnum2[idx2])-ord('0') carry,rem=divmod(n2*n1+carry,10) a+=rem*(10**(idx1+idx2)) if carry: a+=carry*(10**(idx1+idx2+1)) return str(a)
Author(s): Liu J, Grundy SM, Wang W, Smith SC Jr, Vega GL BACKGROUND: The relative contributions of the metabolic syndrome (MetS) and dysglycemia on the risk of cardiovascular disease (CVD) have not been dissected. We aimed to compare MetS with dysglycemia in their association with the 10-year incidence risk of CVD. METHODS: A total of 30,378 subjects were recruited from 11 provinces in the CMCS and followed-up for new coronary heart disease (CHD) and stroke events (ischemic stroke and hemorrhagic stroke) for 10 years. Incidence rates and HRs were estimated by the presence or absence of MetS, impaired fasting glucose (IFG) and diabetes, and by the various traits of MetS. RESULTS: Among the subjects, 18.2% were defined as having MetS; 21.1% had IFG, and 6.8% had diabetes. Metabolic syndrome prevalence in IFG and diabetes was 38.1% and 48.7%, respectively, and the prevalence of IFG and diabetes in MetS was 44.1% and 18.3%, respectively. After adjusting for nonmetabolic risk factors, HRs of total CVD, CHD, and ischemic stroke in MetS were significant and higher than those in non-MetS, regardless of glycemic status. In the absence of MetS, the impact of dysglycemia was found only in IFG to CHD and diabetes to ischemic stroke. Hyperglycemia without any concomitant disorders was not associated with significantly higher risk of CVD. CONCLUSIONS: The increased CVD risk in individuals with IFG or diabetes was largely driven by the coexistence of multiple metabolic disorders rather than hyperglycemia per se. Identification of clustering of metabolic abnormalities should be given more consideration in CVD prevention.
Heredity or Environment Charles Horton Cooley University of Michigan I take it that the formulation of a scientific problem is designed to promote truth by summing up what has been ascertained and indicating where more illumination is needed. For the past fifty years or thereabouts (since Galton began to write on Nature and Nurture) it has been usual to accept the heredity-environment dilemma as a formula of this sort and to endeavor to advance knowledge by extending one horn of it or the other. I have long been dissatisfied with this discussion, and the more I think of the formula in question the more it seems to me illusive, out-of-date and tending "t present rather to obstruct and misguide the pursuit of truth than to further it. It does not sum up the known facts but puts in their place a quasi-metaphysical antithesis which has no precise meaning and suggests antagonism between two aspects of what should be seer as an organic whole. Let us look at its terms. Heredity is something carried. by the germ-plasm, which disappears, as such, when conception %as taken place and gives way to the developing organism. Precisely speaking it does not exist as a concrete fact in the later life of an individual. It is an origin, a historical antecedent, which is now merged and lost in a new phenomenon, namely development. Like the Battle of Gettysburg it has immense historical importance — ( 304) no doubt we should have been very different if it had gone otherwise—but no contemporary reality. Therefore it is untrue to say that heredity is one of two factors to which we must ascribe what is now taking place. The immediate factor is the organism, which although the successor of heredity is by no means the same thing. And what is environment? Apparently an external and surrounding condition, distinguished from an organism, which is a living thing that it surrounds. It is easy to see the fish, for example, in his environment the water. But does the distinction represent for scientific uses an actual state of facts? In the lower forms of life, whose relations with their surroundings are simple, it perhaps does. But as we go upward in the scale to human life, and particularly when we come to civilized man, we find that organism and environment interpenetrate and mutually absorb each other, so that there is no longer a line of demarcation between them, and the distinction either becomes wholly unreal or requires to be defined in some new way. Are our automobiles and motion pictures part of our organism or of our environment? Certainly of the former, because our life is expressed in them as much as in our words or our thoughts. But they environ us also. On the other hand, the things you might suppose to be wholly in the organism prove to be environment. Social psychologists who wish to preserve the distinction say that anything is environment which is capable of giving a stimulus to the nerve processes. But almost any part of the organism may do this to almost any other part. It means that our stomach, our heart, our ductless glands are environment; also our emotions, memories and thoughts. Our hereditary traits are environment, so that if you regard ( 305) heredity as a fact it is the same fact as its rival. In short, under this conception, (and no other seems to be tenable,) organism and environment are merely aspects of what is, in truth, a single and indivisible life. It would really seem, then, that neither heredity nor environment has any such distinct existence in human life as to make it possible, even in theory, to appraise their several functions. They are, as commonly used, vague expressions not answering to concrete facts that you can see, describe or even imagine; useful currency, perhaps, for some purposes of every-day speech, but not convertible into facts available for science. The conception, as regards human life, is based on the analogy of lower forms, and is indeed a survival from the obsolete biological sociology of a former epoch. But, I ask myself, even if these terms do not stand for separable facts, do they not together cover the whole ground, and must we not assume that one or the other is the reason for everything that takes place in any human life? Let us see. Heredity, we know, is an initial condition, present in the germ-cell, and everything that takes place after conception is development. Heredity and development, then, cover the life process. And development, we have seen, has two aspects, organism and environment. We may diagram this as follows: It would appear, then, that the thing which is alter-native to heredity is development. and not environment, unless you interpret the last to include the organism. Otherwise you leave out the organism. But (and here is the critical point) can we not ( 306) assume that the organism itself is negligible as a cause, and that everything in it must be due to the historical origin (heredity) and the surrounding conditions (environment) ? This is indeed precisely the assumption that the popular statement of the case makes. Notice what it means. It means that we are to set aside the one thing which is present before us as a concrete whole, namely the actual life of the individual, and assume to infer it from a historical antecedent (of which we have in fact no first-hand knowledge, since heredity is always inferred from development) and from a surrounding condition which, as we have seen, has no distinct existence? Is this science? Evidently not; science keeps as close as possible to concrete facts. Far more important to the understanding of an individual life than either heredity or environment, or than both together, even if we could know much more about them than we do, is the study of the actual history of that life, from conception to the present time. "Case study" is the only science of the individual; it includes environment, and if we can learn anything reliable about heredity, let it include that too. Even as a quasi-metaphysical assumption the heredity-environment dilemma is highly questionable. It implies that a life process can be inferred in a mechanical manner from an antecedent life process of a different sort, and from external conditions. Now if there is anything anywhere creative in itself and not mechanically infer-able it is the human life process, and so unless we are willing to abandon the world entirely to mechanism we ought to reserve for this process a certain autonomy. But after all, you may say, are there not real and urgent problems, of race, for example, of genius, of crime, of social classes; problems which in general turn ( 307) upon the question how far social differences are innate and determined by the germ-plasm, and how far they are due to variation in social environments, or perhaps to both causes? Even admitting that the study of the individual is of prime importance must we not also seek to get a general view of these large conditions? If "heredity or environment" is not the right statement what will you substitute? Certainly these are real problems, but what is the heredity-environment formula doing to solve them? Nothing, so far as I know. On the contrary, it tends to keep biologists on the one hand and sociologists on the other fighting straw men and advancing pseudo-scientific proofs that "heredity is more important than environment," or vice versa, which have no effect upon the other side because the two sides have no common ground to start from and the whole game is unreal. A good working idea would first of all,, I should say, show human life as a single organic process in which the germ-plasm, the social process and the various phases of individual development have complementary functions—doing away with the silly notion that the sociologist and the geneticist are natural enemies. It would accept what is definitely known about heredity, glandular secretions, nerve mechanisms, social patterns, group processes and social organization—but not encourage a speculative and particularistic inflation of any of these aspects. It would thus attempt to give some sort of moderation and order to that anarchic extravagance of single track minds by which reason is now obscured. If we must have a slogan let it not be "heredity" nor yet "environment," but "See life whole."
About Ahlburg Sprog - Personal teaching with better results - The way of teaching - The material used for lessons - DP1, DP2, DP3 and Studieprøven - Best way to learn Danish - resources The idea of personal teaching – with better results Back in 2019 Danish tutor Mikkel Ahlburg founded ”Ahlburg Sprog”. Based on his long experience teaching international groups abroad as well as in Denmark, he wanted to offer people from abroad a new teaching method. The method is simple: Personal, engaging teaching – with faster results. “Until Ahlburg Sprog was founded the students could either attend the public language schools and have lessons in big classes with varying levels and changing teachers – or they could pay a fortune to a private company, “ Mikkel says. “In the last case the educators are often students with no or a minimum of teaching experience.” Individual needs – individual motivation required The idea of Ahlburg Sprog was clear: Mikkel saw a need to adapt each lesson to the needs of the individual student, while at the same time being very focused on the student’s motivation to learn. By adapting each online Danish lesson to the individual student, he ensures the motivation remains high. “From my education and long experience in teaching I know that every person learns in his/her own way. Therefore, I know that it is usually very difficult to learn a new language by sitting in a class with twenty other students, “ Mikkel says. “I want to offer each person the best possible way to learn Danish. And the best results are achieved when the lessons are adapted to the level, needs and goals of each student. When the students are motivated and the lessons are flexible, they have time for their full-time job, family, and hobbies.” Current and previous experience: - Owner and teacher of Ahlburg Sprog - Freelance at Berlitz Language Services A/S - Freelance at International Communication A/S - Self-employed teacher for adults Chania, Crete - Teacher at “Den Danske Skole og Kulturforening”, Chania, Crete - Teacher at “Thyregod Skole”, Vejle, Denmark The way of teaching The easiest way to learn and improve a new language is simply by using it. At Ahlburg Sprog we focus a lot on motivating the students to improve their confidence in using the Danish language. We plan each class to make it fit the everyday life of the individual student. During the classes we adjust the lessons to match the student’s motivation, needs and goals. Would you like to: - Improve your small talk skills? - Be ready for the Danish job market? - Comfortably talk to people in stores or in your children’s school? Let Alhburg Sprog help you! The material used for lessons We use the material “Trin For Trin” for the lessons. The material is created specifically for internationals and has plenty of great exercises for reading, listening, and speaking. Furthermore, the textbooks meet the requirements for the different exams at the Danish language schools. The material is carefully selected to make you benefit as much as possible from the lessons. DP1, DP2, DP3 and Studieprøven You may be required to at some point pass one or more of the exams meant for internationals living in Denmark. This could be for one of the following reasons: - You want to join a higher education in Denmark - your work situation requires passing one of the exams - you want to have a family reunification. Passing one of the exams demonstrates your ability to understand, read, write, and speak Danish at a certain level. It furthermore shows that you know how the Danish society works. Result-oriented teaching with the exam as the end goal We get a lot of inquiries from people who want to improve their skills and prepare for one of the exams. We plan our lessons with the scheduled exam as the end goal, and then we focus on the different parts of the exam for the individual lessons. Our experience is that one to one teaching makes for the best results With this way of teaching we can focus on the needs of the individual person and make sure that he or she is ready to pass the exam.
One of the more challenging courses to teach in an engineering discipline is computer programming. Focusing only on algorithms and programming blocks leaves little time to build a connection to daily engineering problems. On the other hand, teaching based on problem solving does not allow for a deeper comprehension of algorithm development. “Finding the right balance between these approaches was the issue that we had faced in the Cal Poly Pomona Aerospace Engineering Department for teaching MATLAB,” Dr. Nakhjiri explains. “I decided to try a different approach.” A traditional textbook cannot adequately engage the students in the process of learning. “No textbook could provide the level of student interaction I was hoping for, until I found the MATLAB zyBook,” he notes. The zyBook platform not only encourages the student to become an active learner; it also challenges the traditional learning environment with a focus on hands-on experience. “Cal Poly Pomona Aerospace Engineering is famous for its learn-by-doing approach. Why not use an innovative platform such as zyBooks to implement the same approach to teaching MATLAB programming?” he suggests. Dr. Nakhjiri used zyBooks’ platform for a relatively large group of students in a flipped classroom style. Students completed readings and embedded activities in the MATLAB zyBook each week, and lecture time was dedicated to applying that knowledge to simple engineering problems. “A flipped classroom provided the opportunity to discuss engineering problems with my students knowing that they have already learned the material needed for that week,” reports Dr. Nakhjiri. The interactive segments in the zyBook promoted active learning, and the instructor dashboard allowed Dr. Nakhjiri to easily track student progress and identify common challenges to discuss in class. Using these tools, he could determine which students struggled with a certain topic and provide them with appropriate resources. This approach would be much more difficult with a traditional textbook. Student feedback at the end of the quarter was positive. They liked the zyBook and found it very effective compared to traditional textbooks. Many students showed more interest in studying other programming languages. For Dr. Nakhjiri, affordability, excellent customer service, and meaningful student engagement are his reasons to continue using his zyBook. He concludes: “The interactive zyBook platform is a step forward toward personalized education. It is now upon us, instructors and students, to help improve these platforms by sharing our experience and needs with innovators. I look forward to using the MATLAB zyBook next year in my class.” Dr. Navid Nakhjiri is an assistant professor of Aerospace Engineering at Cal Poly Pomona. He is passionate about exploring new ideas and teaching methods that could result in a more productive classroom experience for his students. Dr. Nakhjiri believes that the future of education will include more personalized teaching approaches that meet each student’s unique strengths and abilities.
Dry Weather and Weed Control Posted: July 13, 2012 Some crops are still being managed for weeds. However, with much of the region in drought, there is some concern about the success in trying to control certain weeds under these dry conditions. Certainly, annual plants become much more “tolerant” with dry weather, maturity, and persistent warm temperatures. “Large-drought stricken annuals are harder to kill”. With perennial weeds, the effect of drought is less clear. Cool season perennials including Canada thistle, quackgrass, and dandelion will certainly go into a summer dormancy period when dry warm weather persists. If possible, they should not be treated with an herbicide until actively growing. Cool season perennials mimic the same growth cycle as your lawn; active in the late spring and early summer followed by a slow period and then a rebound in later summer and early fall. Once the “heat” of summer has passed and assuming they have relatively healthy green leaves, then an effective systemic herbicide should work well. Some things to consider as we experience dry weather and the weeds continue to grow: - Increase the herbicide rate if the label allows and make applications at the most favorable time for increased control. Make applications in the morning when the weeds are most active. - Apply herbicides to smaller weeds or wait a few days to spray if rainfall is in the forecast. - The post grass herbicides (Assure, Select, Poast, etc.) tend to be one of the most susceptible groups to decreased efficacy in dry conditions followed by the ALS-inhibitors (Sandea, Matrix, Raptor, etc.). Contact herbicides (Reflex, Liberty, Gramoxone, etc.) are generally less effected by drought stress, but be sure to increase carrier volume to achieve good coverage. - Think about adjuvants. You may need to use a higher rate or switch to MSO (methylated seed oil) or COC (crop oil concentrate) if they are allowed which can increase herbicide uptake and improve control. However, remember oil-based adjuvants can also increase the potential for crop injury. Sometimes there is a fine-line between controlling the weeds and injuring the crop.