Tuesday, December 24, 2019

Why Schools Are Not to Blame for Cyber Bullying Essay

A young boy was once sitting on his computer when he saw an email from one of his peers. He opened the email, and soon enough realized he was being bullied. However, the young boy let the cyber-bullying get out of control. His parents told the school about it, but there was nothing they could do. The boy had to go to a new school where he wouldn’t be bullied. There are many problems like this that get out of hand, so much that the school can do nothing to help. Many people blame the school for not helping, when there is nothing they can do about it. Therefore, schools do not have the responsibility to respond to and protect students from the challenge of cyber-bullying, even if it’s off campus. Schools do not have the responsibility to†¦show more content†¦They said that students must learn themselves that they should not say something online that they wouldnt say in person. For example, if a student calls someone ugly online, they probably wouldnt say that in person. However, the school cannot do anything to stop this because they did it at home.The school systems cannot be held responsible for things that students do at home, and they should therefore, not be responsible to take action against cyber-bullying. Others would say that schools do have the responsibility to prevent bullying when the schools helping to prevent bullying empowers students to face up to their bullies. However, this is not the case.Students would not be empowered to stand up to their bullies if schools took responsibility for the students. Instead of standing up to the bullies themselves, the students would probably try to make the school take care of it. This goes to show that having the school take care of bullying for the students, would not help the students learn how to take care of things on their own. For example, if the students figured out that the school would take care of something for them, they would not want to do things themselves. Most students would probably get lazy. Therefore, schools should not be bound by law to take care of bullying because students would become lazy. Lastly, schools do not have the responsibility to respond to the cyber-bullying of their students is because sometimes, students do not tellShow MoreRelatedHow is Cyber Bullying Done? Essay621 Words   |  3 Pages Think about it? How is Cyber Bullying Done? To find an answer to this question is hard. Cyber bullying is done in many different ways around the world. Cyber bullying can happen anywhere at any time, even in places where you normally feel safe, such as home. Allot of cyber bullying can be done anonymously, you may not be sure who is targeting you. Since cyber bullies cant see your reaction, they will often go much further in their harassment or ridicule then they would face to face with youRead MoreCyber Bullying Must Be Improved And Consistent1267 Words   |  6 Pages Amanda Bridges Ms. Shultz / Mrs. Seymore English 5th Period 28 January 2015 Cyber Bullying Have you ever been cyber bullied? The effects of cyberbullying are greater than many people have come to realize. Cyber bullies have been around for a long time, but technology now gives them a whole new way to get to their victims easier and faster. Cyberbullying is the use of information technology to repeatedly harm or harass other people in a deliberate manner (abouthealth). Cyberbullying happensRead MoreParents Should Not Be Blamed For Bullying1075 Words   |  5 Pages Have you considered why we blame parents for their children being bullied? Parents should not be blamed for something their child did. I’ve researched this topic for awhile. We all should know that bullying is a major issue in today’s society. What’s even worse is the parents are being blamed for this! Though it was the bullies fault, the parents still get in trouble. This argument will prove that parents shouldn’t be BLAMED for bullying someone when it was their child; parents do n’t have twenty-fourRead MoreThe Social Problem Of Peer Pressure1659 Words   |  7 Pagesin Los Angeles, California who are peer pressured into bullying by their peers and conforming to peer pressure that leads to Cyber-bullying and Cyber-crimes. Bullies are misinterpreted as perpetrators; however, they were at some point victims too and created due to lack of social intervention and significantly affected by the victimization by their peers. Cyberbullying: Through the eyes of children and young people, it acknowledges how bullying has entered a â€Å"digital era† and the need for professionalsRead MoreCyberbullying Is Growing At An Alarming Rate1222 Words   |  5 Pagesdrugs, skip school, recieve poor grades, and have low self esteem. These negative effects of cyberbullying can get much worse; according to a study conducted by Yale University, 7% to 9% of suicides are committed due to some sort of bullying. Since the popularity of social media is increasing, we could expect that percentage to rise. We can no longer ignore cyber bullying because of its negative effects’, like suicide, and we should implement programs to reduce it and protect school age studentsRead MoreCyber Bullying Essay example1112 Words   |  5 Pageshas taken bullying to the extreme. Before advanced technology, children and teenagers were troubled by school bullies. Today, they are targeted by bullies via internet. Most people under go some form of bullying in life. Growing up, I experienced the typical name calling and spiteful comments. I also dealt with minor cases of virtual bullies. These bullies are known as cyber bullies. I find this topic compelling because it’s a matter that has spiraled out of control. â€Å"Cyber bullying is definedRead MorePublic Schools Should Be Able to Regulate Bullying Essay example1356 Words   |  6 Pagesteasing. When one reads these words, the first thing that we might associate it with is bullying. Whether or not we have experienced it ourselves, we have at least heard about it. However, as technology advances, so does the method of bullying. Today in our world that is run by technology, the modern method of bullying is called cyber bullying. The most recent definition of this is, â€Å"†¦Ã¢â‚¬ ¦Ã¢â‚¬ ¦Ã¢â‚¬ ¦Ã¢â‚¬ ¦Ã¢â‚¬ ¦Ã¢â‚¬ ¦This type of bullying allows the perpetrator to hide behind a computer screen or a phone screen and harass theirRead MoreCrt 205 Final Paper1576 Words   |  7 Pagesout one form for each source. |Source 1 Title and Citation: Cyberbullying Has a Broader Impact than Traditional Bullying | |Uhls, Yalda T. Cyberbullying Has a Broader Impact than Traditional Bullying. Cyberbullying. Ed. Louise I. Gerdes. Detroit: | |Greenhaven Press, 2012. At Issue. Rpt. from Is Bullying Going Digital? Cyber Bullying Facts. PsychologyinAction.org. 2010. Gale | |Opposing Viewpoints In Context. Web. 23 May 2012. Read MoreThe Effects Of Cyberbullying On Children And Teens1568 Words   |  7 Pagesbecoming more of an issue, some schools around the world have been enforcing rules to stop cyberbullying. 88% teens who use social media have seen someone be disrespectful to another individual. â€Å"It is no secret that teens are now spending more time than ever on social networking sites and due to this increase many are using it to bully their peers†. According to Sherri Gordon who wrote the article â€Å"Reasons Why Kids Cyberbully Others† here are some reasons why. Some reasons that children cyberbullyRead MoreThe Psychology behind Bullying in Adolescents Essay1176 Words   |  5 PagesBullying by definition is a form of aggressive behavior in which someone intentionally and repeatedly causes another person injury or discomfort. Bullying can take the form of physical contact, words or more subtle actions. [Why] Bullying has grown new outlets over the last decade. With social media outlets and text messaging added to the game, bullying is not just about getting tormented face to face anymore. Cyber bullying can include sending out mean or threatening emails and instant messages

Monday, December 16, 2019

Consumer Behavior †Chapter 1 Free Essays

Consumer behavior chapter 1 – introduction In our ever-increasing global market, attention needs to be paid to the global consumer in order to identify who they are. Marketers and consumer behaviour researchers are constantly attempting to determine the similarities and the differences of the multifaceted global consumer. Some general theory about the sociological and psychological influences on consumer behaviour may be common to all Western cultures, but there are still strong We will write a custom essay sample on Consumer Behavior – Chapter 1 or any similar topic only for you Order Now com/pricing-psychology-consumption/"regional consumption trends. Consumer behaviour as it pertains to the marketplace is concerned with the manner in which consumers purchase and use products and services, and how these goods are incorporated into their lives. Researchers use various techniques to determine the features of consumers. These techniques include categorisation by the use of descriptive characteristics, such as demographics and psychographics. Individual consumers are part of a larger society with common cultural values and beliefs, which can be further segmented into smaller groups, or subcultures. There are regional differences between consumers from different European countries, despite their being part of the larger European community. Marketers use  market segmentation strategies  to adapt their communications to meet the varied needs of a diverse target market. Consumer behaviour is good business! Marketers who understand their customers’ needs and wants do better than their competitors. The essential component is deciding who to target and how. Marketers do this by defining markets and then dividing these markets into smaller chunks, or segments. A segment is based upon the demographic and psychographic aspects of the targeted population. The consumer’s response is the ultimate test of whether or not a marketing strategy will succeed. However, initial consumer testing is not an absolute science. Sometimes a product can fail its initial test marketing and still go on to be highly successful. Sony has responded to market and consumer changes with its latest MP3 player, a rival to the Apple iPod. In attempting to define accurate information about the traits, habits, likes and dislikes of consumers, effective market segmentation separates members of a population into smaller segments of consumers who share unique characteristics that set them apart from the main population. Market segmentation  can be described as ‘slicing up’ the market into bit-sized chunks of information. It is essential that for effective segmentation the following criteria are met: Consumers within the identified segment share similar product needs that will be different to other consumers in other segments. Important differences between the segments can be clearly identified. The segment is large enough to be profitable. Consumers in the segment can be reached with an appropriate marketing mix. The consumers in the segment will respond in the desired way to the marketing mix. Two important segmentation tools are the use of demographics and psychographics to determine unique attributes of the targeted population. Demographic statistics measure the observable aspects of the population. Psychographics are far more subtle. They illustrate the differences in consumers’ personalities and cannot be measured objectively. Demographic variables include: Age Different age groups have different needs and wants. Consumers in similar age groups often share similar values and cultural experiences. Gender The behaviours and tastes of men and women are constantly evolving. It is thought that gender segmentation is an unintended result of marketing strategy. Family structure Marital status and family are important demographic variables that often affect spending priorities. Social class and income People in the same social class often have similar occupations and may share values and interests. Income is a good determinant of buying power and market potential. Race and ethnicity Immigrants from Asia and Africa are the fastest growing ethnic groups in Europe. Increased multiculturalism creates new opportunities for marketers. Geography Cultural differences play an important role in consumption patterns. Within the European context, consumption is common across cultures, but it is also specific between different cultural groups. European segmentation factors include: Consumption traits that are common across different cultures. Consumption traits that are specific to cultural groups. Relationship Marketing: building bonds with consumers The more marketers understand their current and potential consumers, the more effective they will be. This concept has developed into the role of  relationship marketing, in which marketers build long-term relationships between the company and the consumer. Relationship marketing is seen as the big trend for marketers. In conjunction with re lationship marketing, a major growth area has been the need to gather and manage accurate information about customers. This has culminated in the formation of sophisticated  database marketing  that tracks consumer buying habits. However, this ‘big brother is watching you’ concept intimidates some consumers. Database marketing is required to comply with respondent confidentiality restrictions. As global consumers, we are influenced significantly by marketers through a number of mechanisms centred upon building relationships with customers. Advertisements illustrate how we should act and define roles, as well as informing about the use of products. As of 2006, the majority of people will live in urban centres and the number of ‘megacities’ will increase to 26 in 2015. A controversial effect of the sophisticated marketing strategies used to communicate with these large potential markets is the emergence of  global consumer culture, in which people are united by their common devotion to brand-name consumer products, film stars and music celebrities. 2 Virtual consumption and the power of crowds There is little doubt that the digital revolution is one of the most significant influences on consumer behaviour, and the impact of the Web will continue to expand as more and more people around the world log on. It is not all bout businesses selling to customers (B2C e-commerce). The cyberspace explosion has caused a revolution in consumer-to-consumer activity (C2C e-commerce): welcome to the new world of  virtual brand communities. A close cousin of the global consumer is the  Euro-consumer. Much research and marketing effort is focused upon determining the characteristics of this consumer. Findings indicate links between social class, culture and income, with consumers who expressed more positive attitudes towards cultural change being more likely to consume luxury goods, despite their demographics and social class. 2 Marketing Ethics There is a constant battle between achievement of organisational profit goals and the necessity to conduct business honestly and ethically. 1 Prescribing Ethical Standards of Conduct Organisations often devise codes of conduct for their employees, and marketing associations provide guidelines for conduct, such as the following: Disclosure of the substantial risks associated with a product or service. Identification of added features that will increase the cost. Avoidance of false or misleading advertising. Rejection of high-pressure selling. Prohibition of selling or fund-raising under the guise of market research. Socially Responsible Behaviour Some marketers continue to violate the bond of trust with the consumer with the use of ‘bait-and-switch’ tactics and misleading claims. A critical barometer of ethical behaviour is the action taken by companies to rectify a problem with their advertising or products. Benetton discontinued their controversial advertisements that f eatured the depiction of an AIDS victim and which were deemed to be exploitative by consumers. Much criticism has been leveled at marketing techniques that influence consumers to purchase products that are not essential. Philosophers view this conundrum from the perspective of individuals exercising their free will and free action in coming to decisions. However, this clinical approach supports the argument so long as only informative advertising is used in the marketing communication. Any other type of advertising – such as ones with imagery and underlying motives – cannot be so clinically rationalized. The raging debate that marketing techniques manipulate customers continues. We explore this issue from three perspectives. 3 Do Marketers Create Artificial Needs? Some conservative traditionalists believe that advertising can lead to the moral decay of societies by the pursuit of hedonistic pleasure, while some liberals believe that the misleading promises of material pleasure function to buy off people who would otherwise be revolutionaries working to change the system. Through advertising, then, the system creates demands that only its products can satisfy. Numerous responses to this criticism abound, including the argument that needs are a basic biological motive and that a want represents one way that society has taught us that the need can be satisfied. The primary objective of advertising is to create awareness of needs, rather than to create the needs themselves. Needs are thus something that we are socialized to have. Marketers must be cognisant of the fact that they contribute heavily to the socializing influence of individuals in modern society and, thus, influence needs. Some commentators believe that purchasing decisions are manipulated by marketers using their understanding of psychology and the social sciences. They claim that marketers devise advertising techniques that manipulate a majority of individuals. A sense of social responsibility should underpin all advertising. There is little doubt that advertising creates and changes patterns of consumer behaviour. The central issue is whether the audience is discerning enough to see through the marketing hype and still make a deliberative judgement. Have a look at the  PBS Frontline website on ‘The Persuaders’, which explores the world of modern marketers and advertisers, and discusses the methods they use to decipher who we are and what we want. As powerful an influence as advertising is, in the majority of cases advertisers still do not know enough about their target audience to effectively manipulate them. This fact is supported by the high degree of product failures (40-80 per cent) that occur. 4 Interdisciplinary Influences on the Study of Consumer Behaviour Consumer behaviour is a multidisciplinary field that integrates a vast array of disciplines, from psychophysiology to sociology and anthropology. Each discipline views the subject from their own perspective, which results in a rich sociological interpretation of a fascinating and dynamic topic. The figure below illustrates the integrated and interconnected approach to consumer behaviour from both a macro and micro perspective. The macro perspective represents the social focus and the micro perspective represents the individual focus. [pic] pyramid of comsumer behavior Consumer research has adopted two popular forms of research paradigms, or sets of beliefs: Positivism Believes that human reason is supreme and that objective truth is only revealed by science. Structured and ordered. Stresses functionality. Celebrates technology. Regards the world as an ordered, rational place with a clearly defined past, present and future. Emphasizes material well-being, and is based predominantly on the homogenous views of a predominantly Western and male culture. Interpretivism Believes the social and cultural world in which we live is complex. Dislikes too much emphasis on science and technology. Emphasizes the importance of symbolic and subjective experiences. Believes individuals construct meanings based on unique, shared cultural experiences, therefore there are no right or wrong references. Consumption is interpreted as a set of diverse experiences. Many consumer behaviour experts believe that the existing predominant positivistic paradigm of consumer behaviour is undergoing a paradigm shift, due to challenge from the competing paradigm of interpretivism. How to cite Consumer Behavior – Chapter 1, Papers

Sunday, December 8, 2019

Does a dividend policy matter free essay sample

That report is a detailed review of dividend policy and whether or not could affect the market value of the company. When companies make profits, managers have to decide either to reinvest those profits for the good of company or either they could pay out the owners (shareholders) of the firm in dividends. Once they decide to pay dividends they may possibly establish a permanent dividend policy, which is the set of guidelines a company uses in order to decide how much of its profits it will pay out to shareholders in dividends and that decision depends on the preferences of existing and new investors and the situation of the company now and in the future (Garrison, 1999). There are various limitations that may affect firm’s decision and must consider when paying dividends to shareholders such as Legal Limitations where added net realised profits is the only way to pay dividends, Liquidity where managers has to consider the effect that future dividend payments may have on liquidity, Interest Payment obligations where if the gearing (level of debt) is high then the available funds for dividends should be reduced and Investment Opportunities where a company could invest in attractive projects rather than to pay more dividends (Watson and Head, 2007). There are two theories related to dividend, the Irrelevance Theory suggested that dividend policy it is not relevant to security valuation and the Relevance Theory, suggested that it is relevant and affect the value (Bar-Yosef and Kolodny, 1976). Below will see those theories. IRRELEVANCE THEORY Modigliani Miller (1961) through the Irrelevance Theory stated that share value depends on corporate earnings, which reflect the investment policy of the company, depends only on investment decisions and it is independent of the level of dividend paid. First of all that theory assume that capital markets are perfect, there are no transactions costs associated with converting shares into cash by selling them and firms can issue shares without incurring flotation or transactions costs to raise equity, whenever needed (Damodaran [Internet source]). Another assumption that Modigliani Miller made is that there are no taxes at either a corporate or personal level associated with dividend and informations are freely available to all investors. Continuing with the assumptions Modigliani Miller stated that In a perfect capital market there are no conflicts of interests between managers and security holders, which is known as the Agency Problem. Shareholders, actually, own a company but managers are the ones who make the business run and decide. The agency problem arise because manager’s interests are different from shareholder’s interests and that is for the reason that managers may prefer to invest in unprofitable projects for their own benefit and that may incur some costs in order to manage the manager’s behaviour (Ming and Ming, 2013). Modigliani and Miller argues that rational investors, in other words those who prefer their wealth maximization, do not care whether they receive dividends on their shares or investing retained earnings in new opportunities, they have identical borrowing and lending rates and were apathetic to the timing of dividends. Furthermore shareholders can simply sell some of their shares for cash, if dividend are too small. According to Modigliani and Miller a company’s choice of dividend policy is a choice of financing strategy and the investment decision is separate from the dividend decision (Watson And Heat, 2007). They also argued that investors calculate the value of companies based on their future earnings capitalized value and is not affected by the dividends that a company pay and neither how dividend policies are set from company. RELEVANCE THEORY On the other hand we have the Relevance Theory of Lintner (1956) and Gordon (1959), who argued that dividends are preferred to capital gains due to their certainty, which means that an investor prefer to receive a certain dividend payment now rather than leaving the equivalent amount in an uncertain investment. If a company pays low dividends may face a fall in share price and that is because investors exchange their shares with shares of a different company with higher dividend policy. Under that theory we have Asymmetry of Information, which means that dividend decisions may contain new information for shareholders and that is because managers have more informations about the health of the company than investors. Asymmetry of information arises when capital markets are not perfect and depend on the direction of the dividend change and the difference between the actual dividend and the expected dividend by the market. Lintner Gordon argues that shareholders are not homogeneous, they have different needs and preferences and the majority of them need a fixed income preferring dividends to capital gains which depends on their personal tax circumstances. A company’s share price is affected downward form the disappointment of its shareholders if there is a significant change in its dividend policy. Lintner and Gordon use a mathematical model, knows as the Dividend Growth Model, to predict the value of ordinary shares through an increasing stream of cash flows. Dividend growth model is the equation:Po=(Do(1+g))/(r-g)=D1/(r-g), where: Po is the current market price of the share, D1 is the dividend at time t1, g the expected future growth rate of dividend, r the required rate of shareholders and Do the current dividend. The model shows the relationship between the payout ratio, the rate of return, market price of the share and the cost of capital (Answers, 2013). Both theories are established and argued by famous economists and we cannot challenge them, but under a real market conditions we cannot use the Irrelevance theory and that is because some of the assumptions made by Miller and Modigliani are not realistic. Transaction costs are not zero hence there is a price for investors who try to sell their shares and neither informations are freely available to all investors. Taxes does exists in the real world and issuing securities does incur costs. Moreover the assumption that investors have free informations available for them is unrealistic such as they have to spend time and money in order to have those informations (Watson and Head, 2007). For the reason that Irrelevance Theory’s assumptions are realistic under a perfect market only, the Relevance Theory is more useful in real world since a firm’s valuation is affected from the distribution of cash to investors (Ming J Ming X, 2013). POLICIES Using the more realistic theory there are some policies that should be used based on two factors: the company’s operation industry and second the nature of the company and its characteristics. As will see below those policies have advantages and disadvantages. Those Policies are: The Fixed percentage payout ratio policy, The Zero dividend policy and The Constant or steadily increasing dividend Fixed Percentage Pay-out Ratio is the policy where the ‘company pays out a fixed percentage of annual profits as dividends’ (Watson and Heat, 2007). In that case, firms can create reserves for the years that earnings are fewer than usual or when they have losses. Is better for companies with stable earnings over years (SlideShare, 2012). The main advantage is that a company who choose that policy can easily control and send clear signals to its shareholders, related to the level of company’s performance. The disadvantage is from the company’s point again and is that it limits the available funds for reinvestment (Watson and Heat, 2007). In Zero Dividend Policy ‘a company could decide to pay no dividend at all’ (Watson and Heat, 2007). That policy is more acceptable for new companies or for firms which need those profits for research new possibilities in the company. However is unacceptable to the most shareholders and that is because the majority of investors are rely on dividend payments as an income and that is the disadvantage of that policy. On the other hand the advantages are that is by not paying dividends, the administration costs will be eliminated making the operation easier and most important all the profits of the company could be reinvested making the company more attractive for new investors (Watson and Heat, 2007). In Constant or Steadily Increasing Dividend Policy, a company may choose to pay dividend in a constant or steadily increasing ratio. In that case investors prefer a yearly steady growth of dividends and avoid to invest in companies with fluctuating dividend (Lee, 2009). That policy counter to the zero dividend policy is more acceptable from the majority of investors and the reason is that they expect that payments will continued for ever. That except of the advantage may cause one disadvantage, the problem that shareholders expect dividends that company may be unable to afford. Another disadvantage is the fact that companies who adopt that policy do not have the ability to invest in projects that may be profitable (Watson and Heat, 2007) CONCLUSION Concluding this report we can say in sure that companies is likely to prefer the theory of Lintner and Gordon, the Relevance Theory, which argues that dividend policy does matter in real world and in the structure of the real market. Then, according to the policies stated above, should be able to choose which dividend policy is better for them. We hope that the contents of this report are useful and will help you to find the best for the company. We are waiting for your chief’s accountant comments and we are ready to help out you on any further matters which may arise on these or any other finance issues.

Saturday, November 30, 2019

Puryfying Used Cooking Oil free essay sample

The researchers are trying to figure out the effects of sedimentation, activated carbon, and decantation and boiling on purifying used coconut, palm and vegetable oil. The researcher’s experiment resulted to the change of appearance, odour and viscosity of each type of oil. The now purified cooking oil is faster to heat which makes cooking easier, faster and more efficient. The purified oil is quite beneficial however it does not take in as much taste as the unpurified ones. Overall the experiment was very successful in terms of finding the positive differences in each type of oils. The vegetable oil was the best product of all the processes because it had shown the best improvement in all aspects including appearance, odour and viscosity compared to the coconut and the palm oil in the experiment. This study can benefit people who love to cook. Acknowledgements The researchers would like to thank the following for making this study successful: †¢The Anico family for openly welcoming the researchers into their home without hesitation. We will write a custom essay sample on Puryfying Used Cooking Oil or any similar topic specifically for you Do Not WasteYour Time HIRE WRITER Only 13.90 / page †¢Ms. Michelle Baldevarona for being patient in helping with the SIP in every step of the way. †¢Most of all, the Heavenly Father for blessing the researchers with minds that are capable of interpreting the information taught and transferring it to useful knowledge. Chapter 1 Introduction Background of the Study In the commercial world of fast food restaurants and Filipino homes, lessening expenses is one their main goals. Most fast food restaurants, such as Jollibee and McDonald’s, try to lessen expenses by reusing cooking oil. More than not, they reuse cooking oil without making sure that it is still sanitary and healthy to use in cooking. Because most Filipino dishes include the use of cooking oil, it is a primary ingredient in many dishes. Therefore, many health concerns are raised, such as increase of cholesterol due to the reused fats present when cooking oil is reused without ensuring its sanitary and nutritional value. Cooking oils undergo a complex series of changes and reactions during heating and frying. Used cooking oils could be purified by removing the odour, undesirable taste and colour substances. Activated carbon, the process of decantation, sedimentation and boiling are potential means of improving the quality of the used edible cooking oils. Statement of the Problem Will sedimentation, activated carbon, boiling and decantation purify coconut, palm and vegetable oil? Which oil is the best product from the purification processes? Hypothesis Purified cooking oil is equitable with unpurified cooking oil in terms of content and quality such as appearance, odour and viscosity. Definition of Terms Activated Carbon is a form of carbon processed to be riddled with small, low-volume pores that increase the surface area available for adsorption or chemical reactions. Coconut Oil an edible oil extracted from the kernel or meat of matured coconuts harvested from the coconut palm Decantation is a process for the separation of mixtures, by removing a top layer of liquid from which a precipitate has settled. Palm Oil – is an edible vegetable oil derived from the mesocarp (reddish pulp) of the fruit of the oil palms Sedimentation – natural process where solid materials sink to the bottom given a period of time Vegetable Oil is a triglyceride extracted from a plant Significance of the Study The study will benefit people who use cooking oil to prepare meals. This will not only save them money but it can also ensure them that their food would still be edible due to the fact that reused cooking oil can easily become rancid (spoiled) and deteriorated to the point it produces undesirable flavours and odours. Besides ruining what would have been a perfectly good meal, rancid oils also contain free radicals that are potentially carcinogenic. Scope and Limitation This study covered the purification of used cooking oils through the use of activated carbon, the process of decantation and of boiling. The researchers used vegetable oil, palm oil and coconut oil to be experimented on. Variables such as the amount of cooking oil used, the length it took to coo, the temperature and food used to cook were controlled. On the other hand, the manipulated variables were the types of cooking oil. Chapter 2 Review of Related Literature Uses and Effects Filipinos are fond of using cooking oil in their homes. They are also conscious of saving money by reusing these oils. But when cooking oils are reused without purifying it, some health hazards may occur. One of these is the formation of 4-hydroxy-trans-2-nonenal (HNE) which is due to the food particles left from the previous food cooked which are reheated again. HNE can cause cardiovascular disease, stroke, various liver disorders, and cancer. Activated Carbon Activated carbon is a form of carbon processed to be riddled with small, low-volume pores that increases its absorption of liquids when passed through it. This can remove the unwanted food particles and further purify it. Sedimentation Sedimentation is the tendency for particles in suspension to settle out of the fluid in which they are entrained, and come to rest against a barrier. Decantation Decantation is a process for the separation of mixtures, by removing a top layer of liquid from which a precipitate has settled. Usually a small amount of solution must be left in the container, and care must be taken to prevent a small amount of precipitate from flowing with the solution out of the container. It is frequently used to purify a liquid by separating it from a suspension of insoluble particles. Coconut Oil Coconut oil is an edible oil extracted from the kernel or meat of matured coconuts harvested from the coconut palm (Cocos nucifera). It has various applications in food, medicine, and industry. Coconut oil is commonly used in cooking, especially for frying and is a common flavor in many South Asian curries. It has been used for cooking (in tropical parts of the world) for thousands of years. Coconut oil is used by movie theatre chains to pop popcorn, adding a large amount of saturated fat in the process. Palm Oil Palm oil (also known as dende oil, from Portuguese) is an edible vegetable oil derived from mesocarp (reddish pulp) of the fruit of the oil palms. Palm oil is naturally reddish in color because of a high beta-carotene content. It is not to be confused with palm kernel oil derived from the kernel of the same fruit, or coconut oil derived from the kernel of the coconut palm (Cocos nucifera). The differences are in color (raw palm kernel oil lacks carotenoids and is not red), and in saturated fat content: Palm mesocarp oil is 41% saturated, while Palm Kernel oil and Coconut oil are 81% and 86% saturated respectively. Vegetable Oil A vegetable oil is a triglyceride extracted from a plant. Such oils have been part of human culture for millennia. The term vegetable oil can be narrowly defined as referring only to substances that are liquid at room temperature, or broadly defined without regard to a substances state of matter at a given temperature. For this reason, vegetable oils that are solid at room temperature are sometimes called vegetable fats. Viscosity The viscosity of a fluid is a measure of its resistance to gradual deformation by shear stress or tensile stress. For liquids, it corresponds to the informal notion of thickness†. Viscosity is due to the friction between neighbouring particles in a fluid that are moving at different velocities. Chapter 3 Methodology Subject of the Study This study made us of three kinds of cooking oil which are mainly used in Filipino homes. These cooking oils are namely vegetable oil, palm oil and coconut oil. Materials ? cup of vegetable oil ? cup of palm oil ? cup of coconut oil Activated carbon Bottle where the decantation process will take place Pot where boiling can happen Procedures 1)After the cooking oil has been used, let the oil stand for a while so that the food particles can settle at the bottom. 2) Pour it through a bottle which in the middle contains activated carbon and has holes at the bottom for the oil to pass through. 3)When youve removed the solids and particulates, pour an amount of water equal to the volume of oil into a large pot or kettle with deep sides. Pour in your oil. Add about 1/2 teaspoon of salt per quart of total liquid to the pot. 4)Bring the oil and water mixture to the boil, and then boil it hard for about 5 to 10 minutes. The darker, more scorched, and/or more strongly flavored the oil, the longer you should boil the mixture. 5)Remove from the heat, and set aside to settle out. It takes about 10 to 30 minutes for the oil to completely separate and come to the top. 6)Pour off the water portion as completely as possible and discard. 7)Put the oil portion back into the deep pot or kettle. Over medium heat, bring to the boil (which for oil is hotter, obviously, than for water. Reduce heat until it is boiling and popping a bit, but slowly. The goal here is to evaporate all of the retained water. When the oil becomes very clear looking and no longer makes any sound (no popping or sizzling sounds), has no more bubbles rising, and no more steam comes to the top, its done. 8)Allow to cool, then bottle in an airtight container for reuse. Chapter 4 Presentation, Analysis and Interpretation of Data Table 1: Observation after Cooking Oils Used and After the Purification Process Type of Cooking OilAppearanceOdour Unpurified Vegetable OilLooks like regular cooking oilContains a subtle scent of something burnt Purified Vegetable OilVery clear yellow colorSmells like original vegetable oil prior to being used Unpurified Palm OilVery dark yellowish-brown colorContains a subtle scent of hotdog Purified Palm OilSlightly lighter shade than previous colorStill contains a subtle scent of hotdog Unpurified Coconut OilMurky yellowish-brown colorContains a subtle scent of hotdog Purified Coconut OilMurky light yellow color Contains a subtle scent that can’t be indentified Major variations were observed before the oils were cooked and after the oils were purified with the processes of sedimentation, activated carbon, decantation and boiling using 6 regular sized hotdogs which were cooked with the temperature between 180 degrees Celsius to 190 degrees Celsius (medium heat) in the span of 5 minutes. Table 2: Viscosity Test Type of Oil1st Trial2nd Trial3rd TrialAverage Unpurified Coconut Oil1. 28 secs0. 98 secs1. 20 secs1. 15 secs Unpurified Palm Oil0. 99 secs1. 15 secs1. 18 secs1. 11 secs Unpurified Vegetable Oil1. 10 secs0. 97 secs1. 15 secs1. 07 secs Purified Coconut Oil 0. 95 secs1. 03 secs0. 91 secs0. 96 secs Purified Palm Oil1. 10 secs0. 93 secs0. 96 secs1. 00 secs Purified Vegetable Oil0. 97 secs1. 06 secs0. 94 secs 0. 99 secs The test was done with 100ml of different cooking oils, both purified and unpurified. The weight dropped into each of the container weighed 50grams. The most viscous liquid was the unpurified cooking oils and the unpurified was the lesser viscous one. The viscosity of the cooking oils affects how fast it takes to heat it. The more viscous a substance is, the longer it will it take to heat it. Although less viscous oils are faster to heat, foods cooked in it do not take in as much as the taste of the oil compared to the more viscous one. The purified oils were less viscous due to the purification processes done with it. Chapter 5 Conclusion and Recommendation Summary The outcome of the researchers’ purification of the cooking oils deemed to be successful. These positive differences were proven in the observation table and viscosity test. The three oils used, mainly coconut, palm and vegetable, all showed these differences in their appearance, odour, and viscosity. Conclusion Our experiment proved to be successful in terms of finding key differences in both the used and purified cooking oils. Based on the tables presented; there were significant differences between the unpurified cooking oil and the purified cooking oil in terms of its colour, odour and viscosity. The purified cooking oils; however, proved to be more beneficial in those different aspects. But overall, it was the vegetable oil that was the best product of our purification process because it had the most improvement in appearance, odour and had the lesser change in viscosity compared to the other two. Recommendations For future use and investigation, the researches recommend that a wider variety of oils, such as olive, canola and peanut, should be tested. The researchers also recommend that more purification processes should be tested with the oils to further purify it. And lastly, a nutritionist is recommended to tests the nutritive value of these oils whether or not they have changed after being used and after purifying it.

Tuesday, November 26, 2019

The Distance Between Degrees of Latitude and Longitude

The Distance Between Degrees of Latitude and Longitude What is the precise location of Los Angeles? It can be stated in relative terms (about 3,000 miles west of New York, for example), but for a cartographer, pilot, geologist, or geographer, a much more specific measurement is needed. In order to precisely locate any spot in the world, therefore, we use a geographic coordinate system that is measured in degrees of latitude and longitude. This system starts with an imaginary grid of lines that cover the entire planet. Locations are measured based on both X and Y coordinates within the grid. Because the Earth is round, however, the distances between lines on the grid vary. Defining Latitude and Longitude Longitude is defined as imaginary lines called meridians that run from the north to the south pole. There are a total of 360 meridians. The Prime Meridian, which runs through the Greenwich Observatory in England, is also called the International Date Line. Every location east of this line is one day earlier than every location west of the line. Latitude is defined as imaginary lines called parallels because they are parallel to the equator and to one another. The equator, which runs in a circle around the center of the Earth, divides the planet into north and south hemispheres. Lines of latitude and longitude intersect, creating a grid that allows anyone in any location to pinpoint a geographic location. There are 360 degrees of longitude (because meridians make Great Circles around the globe), and there are 180 degrees of latitude. To further specify exactly where to find anything on Earth, measurements are stated not only in degrees but also in minutes and seconds. Each degree can be broken into 60 minutes, and each minute can be divided into 60 seconds. Any given location can be described in terms of degrees, minutes, and seconds of longitude and latitude. What Is the Distance Between Degrees of Latitude? Degrees of latitude are parallel so, for the most part, the distance between each degree remains constant. However, the Earth is slightly elliptical in shape and that creates a small variation between the degrees as we work our way from the equator to the north and south poles. Each degree of latitude is approximately 69 miles (111 kilometers) apart.At the equator, the distance is 68.703 miles (110.567 kilometers).At the Tropic of Cancer and Tropic of Capricorn (23.5 degrees north and south), the distance is 68.94 miles (110.948 kilometers).At each of the poles, the distance is 69.407 miles (111.699 kilometers). This is rather convenient when you want to know how far it is between each degree, no matter where you are on Earth. All you need to know is that each minute (1/60th of a degree) is approximately one mile. For example, if we were at  40 degrees north,  100 degrees west, we would be on the Nebraska-Kansas border. If we were to go  directly north to  41 degrees north,  100 degrees west, we would have traveled about 69 miles and would now be near Interstate 80. What is the Distance Between Degrees of Longitude? Unlike latitude, the distance between degrees of longitude varies greatly depending upon your location on the planet. They are farthest apart at the equator and converge at the poles. A degree of longitude is widest at the equator with a distance of 69.172 miles (111.321 kilometers).The distance gradually shrinks to zero as they meet at the poles.At 40 degrees north or south, the distance between a degree of longitude is 53 miles (85 kilometers). The line at 40 degrees north runs through the middle of the United States and China, as well as Turkey and Spain. Meanwhile, 40 degrees south is south of Africa, goes through the southern part of Chile and Argentina, and runs almost directly through the center of New Zealand. Calculate the Distance from One Point to Another What if you are given two coordinates for latitude and longitude and you need to know how far it is between the two locations?  You could use what is known as a haversine formula to calculate the distance - but unless you are a whiz at trigonometry, it is not easy. Luckily, in todays digital world, computers can do the math for us. Most interactive map applications will allow you to input GPS coordinates of latitude and longitude and tell you the distance between the two points.  There are a number of latitude/longitude distance calculators available online. The National Hurricane Center has one that is very easy to use. Keep in mind that you can also  find the precise latitude and longitude of a location using a map application. In Google Maps, for example, you can simply click on a location and a pop-up window will give latitude and longitude data to a millionth of a degree. Similarly, if you right-click on a location in MapQuest you will get the latitude and longitude data. Source Latitude/Longitude Distance Calculator. National Hurricane Center and Central Pacific Hurricane Center.

Friday, November 22, 2019

Proton - Definition of Physics Terms

Proton - Definition of Physics Terms A proton is a positively charged particle that resides within the atomic nucleus. The number of protons in the atomic nucleus is what determines the atomic number of an element, as outlined in the periodic table of the elements. The proton has charge 1 (or, alternately, 1.602 x 10-19 Coulombs), the exact opposite of the -1 charge contained by the electron. In mass, however, there is no contest - the protons mass is approximately 1,836 times that of an electron. Discovery of the Proton The proton was discovered by Ernest Rutherford in 1918 (though the concept had been earlier suggested by the work of Eugene Goldstein). The proton was long believed to be an elementary particle until the discovery of quarks. In the quark model, it is now understood that the proton is comprised of two up quarks and one down quark, mediated by gluons in the Standard Model of quantum physics. Proton Details Since the proton is in the atomic nucleus, it is a nucleon. Since it has a spin of -1/2, it is a fermion. Since it is composed of three quarks, it is a triquark baryon, a type of hadron. (As should be clear at this point, physicists really enjoy making categories for particles.) Mass: 938 MeV/c2 1.67 x 10-27 kgCharge: 1 fundamental unit 1.602 x 10-19 CoulombsDiameter: 1.65 x 10-15 m

Thursday, November 21, 2019

Methods & Survey Research Designs Coursework Example | Topics and Well Written Essays - 2000 words

Methods & Survey Research Designs - Coursework Example Positive Relationships: Positive relationship is a relationship that signifies a direct relationship among two variables. That is, when there is an increase in one variable, the other variable is also likely to increase and when one variable decreases, the other also decreases. Negative (Inverse) Relationships: A negative relationship means that increase in the value of one variable leads to decrease in the value of the other variable and vice versa. This relation is also known as an inverse relationship. Pilot Test: A pilot test is a minor version of a large survey test and it is carried out to get an idea of the real test. It involves prior testing of a research tool, for instance, a new information gathering method, and it can also be used to test a hypothesis or design. Critical Theory: It is a social theory aiming toward analyzing and critiquing the society as a whole, in disparity with traditional theory aimed only to explaining it. Critical theories intend to dig under the she ll of social life and expose various theories that render a true and fuller understanding as to how the world works. Cultural Portrait: Cultural Portrait can reflect high moral and spiritual human qualities. It also has the capacity of honestly revealing the negative qualities of the subjects under study. Cultural portraits are mainly common in satirical portraits and caricatures. Bounded System: A bounded system has territories with identifiable edges between the interior and exterior, as well as spaces with different functions happening in different spaces. Examples include an organization, a family, a program or a class in school etc. Discriminate Sampling: It is a procedure which decides the group to which a person belongs according to his or her individual characteristics. Gatekeeper: A gatekeeper in traditional research methodology is a person with whom the researcher has to negotiate entree to participant subjects. The role implies a related position such as, stewardship, own ership or other executive authority along the lines of the presented cultural standards of the research setting. In Vivo Codes: In vivo codes are the factual terms used by researchers mostly as well as the expressions used by various actors also. They tend to be the behaviors which will provide details to the analyst about the methods in which the basic problems of the actors is determined. Memoing: Memoing is the process of recording reflective notes concerning what the analyst learns from the data. Memos accrue as written records or ideas regarding concepts as well as their relations. Progressive-Regressive Method: The progressive regressive method includes a movement in two guidelines. The progressive stage starts with what is clear to examination and the regressive stage returns back to its older roots. Inductive Reasoning: The term inductive reasoning means to analyze starting from bottom up. It takes exact data and creates a broader overview that is considered possible, allowi ng for the information that the end may not be precise. Field Notes: Field notes refer to different notes recorded by researchers throughout or after their study of a specific subject they are engaged in. They are mainly treasured in descriptive sciences that have high implications in this area. Field Journal: The field journal is a notebook that a researcher uses to record personal notes, observational notes, sketches, lists of terms, ideas and so on, when he or she is engaged in a field

Tuesday, November 19, 2019

In what ways did industrialization create new opportunities for women Essay

In what ways did industrialization create new opportunities for women How and why were these opportunities limited - Essay Example Important changes occurred included the elevation of women positions and set up of more demanding jobs. Industrialization shifted American economy from agricultural to an economy characterized by workforce enabling many women to enter the paid jobs. This was possible as women worked in textile industries, mining and agriculture. A change in industrial configuration lessened the number women labors. Opportunities like being employed as a miner were then outlawed by the regime as being illegitimate for woman to toil as a drawer in coal excavating. Another impediment to the employment of women came from gender partition of labor. Gender defined the role of women and women discretely (Hillstrom et al., 205). The cultural devaluing of women household jobs camouflaged its continuation leading to decrease in economy importance. The working place was another cause to limit the opportunities. That was possible when the working places were far away from women homes. Therefore, women were unable to comfortably work and at the same time taking care of their children (Sylvia 2008). Women who got married would rather stay home and look after children due to undue influence from their husbands and core cultural values that need to be

Saturday, November 16, 2019

Savings and Loans Crisis Essay Example for Free

Savings and Loans Crisis Essay INTRODUCTION In the 1980’s, the savings and loan (SL) industry was in turmoil with the watershed event of this being the implementation of price fixing legislation in favour of home ownership in the 1930’s. Even though it was the basis of the crisis, the trigger lies in more fundamental concepts, including fiscal policy, mismanagement of assets and liabilities, pure imprudence by SL institutions, brokered deposits and the cyclicality of the regulation/deregulation process and this was fuelled by economic reactions such as inflation. It would be ‘unfair’ to attribute it to only one factor. Therefore, to properly investigate the crisis and with a view of having all round perspective of the crisis, this report will discuss this financial disaster’s main causes. The impact of the crisis was borne mostly by the SL industry, the savings and commercial banks in the US and more generally, the US economy. This report will further cover the corrective measures undertaken by regulators and the government with the aim of saving the SL sector as the number of institutions with worsening financial conditions steeply increased. The consequences of this crisis persisted until the early 1990’s and this long term effect is understood by analysing the regulations enacted in the aftermath of the crisis. The main turning point has been the enactment of the Financial Institutions Reform, Recovery and Enforcement Act in 1989. Finally, there are essential lessons to be learned from the SL crisis, not only for the SL institutions, but also the banking industry, regulators and the government. CAUSES In the 1930s the SL industry was a conservative residential mortgage sector surrounded by legislation put in place during that period to promote home ownership. At the same time it has its own regulator which is the federal savings and home loan banking loan, and its own insurance firm to insure deposits at SL institutions. However the regulatory and interest rate environment started to change dramatically as from the 1960s when congress applied the Regulation Q to the SL industry by putting a ceiling on the interest rate that SLs can pay to depositors. The purpose was to help thrift institutions to extend interest rate ceiling to them in order to reduce their cost of liabilities and protect them from deposit rate wars since there were inflationary pressures in the middle till late 1960s. Regulation Q was price fixing, and in trying to fix the prices, Regulation Q caused distortion where the costs outweigh any benefits it may have offered. Regulation Q created a cross subsidy, passed from saver to home buyer, that allowed SLs to hold down their interest costs and thereby continue to earn, for a few more years, an apparently adequate interest margin on the fixed-rate mortgages they had at that recent past years. The problem was that the SL industry was not competing effectively for funds with commercial banks and securities market leading to large things in the amount of money available for mortgage lending. The ceiling on interest rate that SL could offer to depositors as per the Regulation Q led dampening of competition for depositors funds between bank and SL. But as new money market funds began to compete fiercely during the 1970s for depositors’ money by offering interest rates set by the market, SLs suffered significantly withdrawal of deposits during periods of high interest rates. This caused outflows from financial institution into higher yielding investment such as capital market instrument, government securities and money market funds. This process is known as disintermediation. Disintermediation has several undesirable consequences. Most important, it both restricted the availability of credit to consumers and increased its cost, particularly for home mortgages, the same consequences affected small and medium sized businesses that did not have access to the commercial paper market. In additional, because normal cash outlays increased to meet deposit withdrawals while cash inflows decreased as new funds were diverted to alternative investments, disintermediation slowed the growth of financial institutions and caused them liquidity problems. To have the cash available to meet withdrawal demands, banks and thrifts were often forced either to borrow money at above-market interest rates or to sell assets, often at a loss from book value. At the same time, rise in oil prices in 1979 pushed inflation and headline interest rates up. Growing inflation in the 1970s received two huge boosts: the first comprised the late-1973 and 1979 oil shocks from OPEC (the Organization of Petroleum Exporting Countries). Soaring oil prices compelled most American businesses to raise their prices as well, with inflationary results. The second boost to inflation came in the form of food harvest failures around the world, which created soaring prices on the world food market. Again, U.S. companies that imported food responded with an inflationary rise in their prices. In order to combat the increase in inflation, there was a rise in interest rates to encourage people to save and spend less. The Federal Reserve opted for tightening monetary measures in reaction to inflationary concerns. As a result of the subsequent monetary tightening, interest rates rose abruptly and significantly. Interest rates soared from 9.06% in June 1979 to 15.2% in March 1980. Such drastic change in base rates caused the yield curve to become inverted. The spread between the 10 year Treasury bond and the 3-month T-Bill became negative as seen in the table below reaching 373 basis points in 1980. (http://www.milkeninstitute.org/pdf/InvrtdYieldCurvesRsrchRprt.pdf) The graph below shows the variation of US Treasury three-month T-Bill. The large rise and the volatility of short term interest rates is evident from the graph. (http://www.milkeninstitute.org/pdf/InvrtdYieldCurvesRsrchRprt.pdf) The following 10-year Treasury against the effective Federal Funds Rate spread also illustrates how the yield curve inverted during the SL crisis. (http://www.milkeninstitute.org/pdf/InvrtdYieldCurvesRsrchRprt.pdf) With high volatility of interest rates during these periods, the SL industry failed to tackle the risk inherent in the funding of long term, fixed mortgages by means of short term deposits. In other words, there was a mismatch of asset/liability with a negative gap and rising short term interest rates. Aftermath In the1982’s, to attempt at resuscitating the SL industry, Congress tried to deal with the crisis by enacting the Depository Institutions Deregulation and Monetary Control Act in 1980 and the Garn-St Germain Depository Institutions Act in 1982, allowed lower capital requirements, which were based largely on book values rather than more market-value oriented techniques, grossly overstate the health of financial institutions. Regulators relaxed regulatory restrictions by decreasing the net worth requirement from 4% to 3% of total deposits, with additional flexibility of not complying with the generally accepted accounting principles (GAAP). The process of deregulation further included the extension for the period of amortisation of supervisory goodwill and the Bank Board removes the specific limitations for the SL shareholders, changing the minimum 400 shareholders restriction to only one, with no one shareholder or group holding more than 10% and 25% respectively and the acceptance of means of payment other than cash. In particular, rules on net worth changed so that thrifts could continue to operate even at historically low levels. Also, â€Å"supervisory goodwill† was used to balance out the books in terms of capital requirements and accounting numbers. This goodwill had no economic sense and simply helped to compensate any institutions, in a merger, when taking over economically impaired assets of insolvent institutions. All in all, the changes in accounting and capital treatment of supervisory goodwill enabled SL’s to post stronger accounting and capital numbers even though the underlying economic situation had deteriorated. There was a cancellation of the ceiling of the loan to value ratio as well. Forbearance or the decline in regulatory oversight was also a major factor of the debacle. Most importantly, savings and loan interest rate ceilings were removed. SL’s had a large proportion of variable rate liabilities (deposits) funding fixed-rate assets they held 84.5% of their assets as home mortgages. These institutions had a negative GAP as the amount of RSL was larger than that of RSA. GAP = RSA RSL Therefore, they were exposed to any rise in interest rates as the yield on the assets were fixed while the cost of liabilities increased. With the rapid increase in base rate in the 1980’s, FI’s cost of RSL rose faster than they could adjust their return on their assets. They had to maintain a high level of interest paid on deposit to avoid deposit withdrawal. The Net Interest Income – the difference between interest on assets and cost of liabilities decreased significantly. NII = Asset Return – Cost of Liabilities On average, the returns on home loans were 9% with an average deposit rate of 7% which implied a 2% net interest income. Given the tight regulations surrounding the SL’s, these institutions relied in the 2% net interest income as their main source of income. However, as the base rate rose dramatically, the NII dropped to negative figures, reaching -1.0% in 1981. Many institutions lost huge amounts of money. Savings and Loans specialised in originating and holding home mortgage loans that were relatively long term assets with fixed interest rates. However, these were funded by relatively short term deposits whose interest rates were variable. There was a maturity mismatch that was exposed to risk of interest rate rise. With the market value of the assets being more volatile because of its longer maturity, and as a consequence a longer duration, the rise in interest rate decreased the value of the mortgages to very low levels. The value of the liabilities decreased as well but to a smaller extent. Since net worth is the difference between market value of assets and market value of liabilities, this led to negative equity of financial institutions. Δ E = (DA DLg) x A x Δr/(1+r) Since DA DLg, with Δr 0, change in net worth value ΔE is negative. The leverage adjusted duration gap between the assets and liabilities was so large and with a large rise in interest rate, the equity value decreased to being negative. By the early 1980s, savings and loans throughout the country were insolvent by about $110 billion and the fund was reporting only $6 billion in reserves (Barth, 1991; Brumbaugh, 1988; Kane, 1989) The legislation also allowed savings and loans to begin to diversify into commercial real estate loans and other loans commercial banks could already make. Congress hoped that if SL’s invested in riskier, and thus, higher yielding assets, they would be able to offset the loss they previously made. The savings and loans were also allowed to originate adjustable-rate home loans. By 1983, most SLs were deemed economically profitable but 9% of the SL industry was insolvent. However, the Federal Home Loan Bank Board (FHLBB) and the Bank Board, went ahead with their plan of regulating the industry by imposing a 7% net worth limit for new entrants in the thrift industry so as to promote safe risk management practices and investments. Although all these developments were intended to help savings and loans, they gave rise to a subsequent twist in the crisis. The new changes did indeed allow savings and loans to reduce their interest rate risks but the changes exposed savings and loans to new risks mainly credit risks. While defaults on the home mortgages were low, defaults and associated losses on other types of loans and investments are typically much higher. By combining interest rate risk with credit risk, spread over a wider geographical area, experienced institutions had greater opportunities to choose a prudent overall balance of risk and return. However, many savings and loans began making commercial real estate loans, activities in which they were relatively inexperienced. Since investing in real estate loans entailed unique risks and required specific skills, SL’s eventually made losses on the real estate loans. These credit quality problems are reflected in the net income of the industry plunging once again, but even more than in the early 1980s, when the yield curve inverted. The industry lost nearly $21 billion in 1987 and 1988, and almost another $8 billion in 1989. Many open but insolvent savings and loans had incentives to take excessive risks and â€Å"gambled for resurrection† because of the phenomenon of moral hazard. If ever something turned wrong, the federal deposit insurance fund would bear the losses; yet the owners would reap the rewards if everything went well. The legislation, however, did not change how premiums were set for federal deposit insurance, meaning that riskier institutions and prudent ones were charged the same premium. Actually, the level of insured deposits was raised from $40,000 to $100,000. The new, lower capital requirements and broader opportunities to lend and invest allowed some savings and loan to take larger risks. With federally insured deposits and the ability to attract more deposits by offering higher rates of interest, deeply troubled savings and loans always had ready access to additional funds. Deregulation encouraged increased risk-taking by SL’s. However, in the mid- to late 1980s, with considerable real estate loans and investments, regional recessions struck the USA, which reduced commercial real estate values. In particular, an unexpected plunge in the price of oil in 1986 contributed to recession. To make matters worse, the Congress passed the Tax Reform Act of 1986 that more than eliminated the tax benefits to commercial real estate ownership it had conveyed only a few years earlier. Commercial real estate values fell dramatically as a result. This severely affected the asset value of the SL’s. In 1987, the Bank Board emphasised the importance of capitalisation by imposing a supervisory approval for SLs which engage in investments that are above 2.5 the multiplier of their tangible capital base. The main turning point was the Financial Institutions Reform Recovery and Enforcement Act (FIRREA), restructuring the industry as a whole by setting up the Resolution Trust Corporation which in total resolved or liquidated 747 thrifts, with assets valued at $394 billion, jettisoning both the FHLBB and FSLIC and setting up a new regulatory institution Office of Thrift Supervision. The key to this act was that instead of trying to save the SLs which were barely solvent, it dissolved them and focused on the solvent ones to reform the industry. With the assistance of market fundamentals – favourable conditions of interest rates, the reinstatement of GAAP accounting and real estate market, the industry stabilised. LESSONS LEARNT The thrift crisis had a bailout plan of $153 billion, of which around 80% was financed by taxpayers. The number of institutions in the SL industry receded considerably until 1995 and before then, the ability of the regulators and the government to deal with the crisis was questioned many times. What followed was a series of court battles, corruption charges and major restructuring. Therefore, consequences were substantial enough for everyone to extract some observations and lessons. The starting point of it all was overregulation, which outlined the restrictions and conditions under which an SL would function. That included rigidity of the institutions to be flexible at a time economic conditions were changing and the financial sector was advancing. With fixed interest rates, it proved difficult for the SL to engage in competition as their means of investing was limited. One crucial point is that additional regulations do not necessarily mean fewer risks. SLs had to assume additional exposure to interest rate risk and alongside with banks, they were prevented from optimising their credit risk exposure. The government sometimes does not modify the regulations as fast as the structure of the industry is changing leading to new risks emerging and the cycle goes on. To keep up with advancement, the government has to put in place tighter risk management policies and controls. However, regulators and government should not direct the investment decisions of institutions. Rather, investments should be in line with market and economic forces. At a later stage, the industry was deregulated in order to remedy the situation. However, this translated into a decrease in market discipline as the SLs made high risk investments as they relied on the safety net of federal guarantee to cover any losses. Moral hazard, adverse selection and passive management were noted. Therefore, it exposes the disadvantage of FSLIC at that time which encouraged the SLs to take long-term and unreported risks. The deregulation, reducing the capital requirements, left the thrift industry more vulnerable to economic changes. From the failure of resuscitating the industry, it was deduced that forbearance treatment towards insolvent institutions might not always be the best option as it can lead to a freeze in lending and stifle the economy. One of the lessons from the thrift crisis which has been consistently taken into account over the years was the reliance on capital ratios. During the deregulation period of the crisis, there was no monitoring of the capital bases of the thrifts which ultimately lead to insolvency. From then on, institutions had to follow certain standard capital requirements put in place by regulators. However, this focus proved recently in the credit crunch to be detrimental, showing that banks favour trust and confidence. It is important to realise that capital ratios and other accounting ratios might not reveal the real economic strength of the institution. The crisis led to more disclosure and market value accounting. It has been understood that it would have been best to restrict involvement of public funds as a means of saving the industry. Using state or public funds to buy-out thrifts below value is not in accord with public welfare. A solution would have been to subdivide the thrifts into insured and uninsured ones with varying degrees of supervisory regulations concerning deposits and investments. One lesson learned was the emergence of an adjustable insurance premium rate which became a function of the institution’s regulatory rating, risk and capital levels. CONCLUSION For some years the final bill for the SL crisis remained uncertain. However, it is known now that, the thrift crisis cost an extraordinary$153 billion – one of the most expensive financial sector crises the world has seen. Of this, the US taxpayer paid out $124 billion while the thrift industry itself paid $29 million. The consequences of the SL crisis for the structure and regulation of the US financial industry were profound. The number of institutions in the SL industry fell by about half between 1986 and 1995, partly due to the closure of around 1,000 institutions by regulators, the most intense series of institution failures since the 1930s. The failures prompted an overhaul of the regulatory structure for US banking and thrifts, a shake-up in the system of deposit insurance and implied Government guarantees. Regulators shifted towards a policy of earlier intervention in failing institutions so that the principal costs are more likely to be borne by shareholders than other stakeholders. There was also a shift towards more risk-sensitive regulatory regimes, with respect to both net worth assessments and the payments to deposit insurance funds, while deposit insurance reform made it less likely that taxpayers would shoulder so great a burden in any future crisis. At a wider level, the SL crisis taught politicians, regulators and bankers how misleading rules-driven regulatory and accounting numbers can be in relation to risky bank activities. At different stages of the crisis, reporting of the financial condition of SLs was deliberately selected by interested parties to cover up the true economic extent of the unfolding disaster. It was a risk reporting failure on grand scale that greatly worsened the long term economic consequences fort the ultimate stakeholder: the US taxpayer. REFERENCES 1. Myth: Carter ruined the economy; Reagan saved it. http://www.huppi.com/kangaroo/L-carterreagan.htm [Accessed 31 October 2010 to 18 November 2010] 2. The U.S. banking debacle of the 1980’s : A lesson in government mismanagement http://www.thefreemanonline.org/featured/the-us-banking-debacle-of-the-1980s-a-lesson-in-government-mismanagement/ [Accessed 31 October 2010 to 18 November 2010] 3. Inverted Yield Curve Research Report, Milken Institute http://www.milkeninstitute.org/pdf/InvrtdYieldCurvesRsrchRprt.pdf [Accessed 31 October 2010 to 18 November 2010 4. The Cost of the Savings and Loans Crisis, FDIC Banking Review http://useconomy.about.com/library/s-and-l-crisis.pdf [Accessed 31 October 2010 to 18 November 2010] 5. The SL Crisis: A Chrono-Bibliography, FDIC http://www.fdic.gov/bank/historical/s%26l/index.html [Accessed 31 October 2010 to 18 November 2010] 6. The Savings and Loan Crisis http://wapedia.mobi/en/Savings_and_loan_crisis.html [Accessed 31 October 2010 to 18 November 2010] 7. US Savings and Loans Crisis, Sungard Bancware Erisk http://www.prmia.org/pdf/Case_Studies/US_SL.pdf [Accessed 31 October 2010 to 18 November 2010] 8. Savings and Loans Crisis, FDIC Report Vol. 1 http://www.fdic.gov/bank/historical/history/167_188.pdf [Accessed 31 October 2010 to 18 November 2010] 9. The Economic Effects of the Savings and Loans Crisis, Congressional Budget Office http://www.cbo.gov/ftpdocs/100xx/doc10073/1992_01_theeconeffectsofthesavings.pdf [Accessed 31 October 2010 to 18 November 2010] 10. The Cost of Savings and Loans Crisis: Truth and Consequences, FDIC Banking Review http://fcx.fdic.gov/bank/analytical/banking/2000dec/brv13n2_2.pdf [Accessed 31 October 2010 to 18 November 2010]

Thursday, November 14, 2019

Free Essays: Comparing Characters and Themes in Hamlet and Macbeth :: comparison compare contrast essays

Parallel Characters and Themes in Hamlet and MacBeth    Throughout William Shakespeare’s plays Hamlet and Macbeth there are many similarities, along with many differences. These plays are both Shakespearean tragedies, which often use supernatural incidents to capture the reader’s interest, and consists of a hero that has a tragic flaw. There are many comparative and contrasting aspects in these plays.   Ã‚  Ã‚  Ã‚  Ã‚  Ã‚  Ã‚  Ã‚  Ã‚  Ã‚  Ã‚   The opening of Hamlet involves a supernatural, as does the opening of Macbeth. In the first scene the ghost of his father, King Hamlet, approaches Hamlet. Similarly, the opening of Macbeth involves the three witches. Although the witches can be seen by anyone they approach, the ghost of King Hamlet is only seen by Hamlet himself, and in one scene by Marcellus and Bernardo, Hamlet’s servants. Similarly in both plays, the main characters are slightly suspicious of the actual powers these supernatural figures have. As the witches use their apparent powers to tell Macbeth the future, the ghost of King Hamlet tells Hamlet what has happened already. Hamlet states in one of his soliloquies â€Å"The spirit that I have seen / may be the devil† (2.2.598-599). Macbeth also has his doubts because when the witches tell him that he will be named Thane of Cawder, Macbeth himself had not known, but many people had. It is possible the witches could have known. In the same matter in both plays, the presentation of the supernatural began to lead to the final downfall of each of the characters. In Macbeth, the three witches cause him to think and do evil deeds. In Hamlet, if he had not seen the ghost of his father, he would not have known that Claudius has killed his father to claim the throne. In both instances the characters gave into the nagging supernatural beliefs. And hence they lost their lives. Other characters in these plays show parallels in their plots. Both plays have a main character that portrays the king of that country. In Hamlet, the King of Denmark, Claudius is directly related to Hamlet. He is his uncle, and also his mother’s new husband. However, in Macbeth the King of Scotland, King Duncan, is not directly related to the main character.

Monday, November 11, 2019

Li & Fung Core Strategies Essay

Li and Fung’s Core Strengths One of the core strengths of Li and Fung was acquisitions as they saw it as a means to sustain growth momentum. Most of the time they tried to â€Å"fill in the mosaic† by acquiring competitors in areas where they lacked positioning, expertise or talent. They just did not acquired companies they made them integrated with the old ones to make When they acquire companies Li & Fung’s strategy is to integrate incoming staff and streamline operations seamlessly within 100 days. Li & Fung works to a Three Year Plan that addresses strategies across the board and in recent years has set and met the goal of doubling its size with each plan. Li & Fung has achieved its milestones a lot faster, while keeping costs down by investing in new technology and software solutions. The company has set up the infrastructure for web conferencing between all of its offices. The company has improved IT software’s, particularly those relative to infrastructure that will offer improved collaboration, such as VOIP phones, WebEx, SharePoint for internal and customer collaboration, videoconferencing and EDI. This has allowed them to cut costs and increase productivity. Also Li & Fung centralized back-office systems contributed to improve the efficiency of acquisition. Li & Fung’s as a middleman provided a wide array of services that spanned the whole supply chain. In addition, they supplied both hard goods and soft goods from suppliers to different countries and territories. Acted as the middlemen, the company could reduce matching and credit risks. How do these core strengths make Li and Fung an attractive option for its customer? How do Li and Fung make the supply chain more responsive (i.e., fast and flexible? Do you think that other companies can replicate Li & Fung’s business model? Why or why not? Be brief.

Saturday, November 9, 2019

How Music Education Helps Students Education Essay

A music instruction is profound on pupils because it provides a assortment of womb-to-tomb wages ; hence, it should be encouraged and instilled in the school course of study.Why Students Should Analyze MusicStudents should analyze music because it offers them the ability to show themselves and larn to work independently and with others every bit good as logical thought, teamwork, perceptual experience, a sense of accomplishment, and an esteem for the humanistic disciplines.Benefits of a Music EducationPersons benefit from a music instruction through as it helps certain parts of the encephalon develop, along with IQ, reading degree, and a sharper memory come with exposure to music. Learning to play an instrument can be compared to larning another linguistic communication and can be disputing at times. Students have to be disciplined in order to get the hang playing an instrument ; it takes clip and committedness which in bend sets criterions of quality that enhances their desire for a cademic excellence.How Music Education Helps StudentsA music instruction allows pupils to show feelings and have a self-identity. Band and orchestra members are less likely to smoke, imbibe intoxicant, or experiment with other drugs than other non-music childs. An instruction in music besides allows pupils to bond with other members in their group who portion common thoughts, beliefs, and behaviours and to be included in multiple coteries, where they feel like they belong.Music Teaches Lifelong SkillsSkills which are indispensable in the work force and community are developed through music survey. The humanistic disciplines besides help excite economic activity by pulling tourers to events and hence increasing the community ‘s local revenue enhancement base. Students involved in their music instruction plan said music is what defines them because it helps them show themselves. A music instruction really does what is said to be true about itA Music Education: The IncomparableIn troductionWhat sort of an instruction do you desire for your kids? Are you satisfied with an instruction sufficient for them to graduate or make you desire them to work at something to the point that it ‘s every bit perfect as they can do it ( Gill, Rogers, and Rogers, 2000 ) ? A music instruction gives pupils subject, strength, and possible along with the resources of teacher support and ego appraisal for every pupil to larn. These accomplishments help pupils all throughout their lives, no affair what they do. The benefits of music instruction are profound on pupils ; hence, it should be encouraged in the school course of study.Why Students Should Analyze MusicWhy should pupils analyze music and is at that place anything society could make to promote all kids to hold the experience? Music offers pupils the ability to show themselves and larn to work independently and with others ( Children ‘s Music Workshop, 2012 ) . Whether it is set or choir, pupils have to pattern th e music over and over by themselves to finally be able to work with others in dry runs and compose an outstanding public presentation. Because of the humanistic disciplines personal connexion to largely everything we want for our kids and schools, they deserve a topographic point in the course of study. Analyzing music besides teaches pupils, logical thought, teamwork, perceptual experience, and esteem for the humanistic disciplines ; plus, it is fun at the same clip and gives pupils a sense of accomplishment which is invaluable to parents and schools ( Frisco School of Music, 2012 ) . In association, with the survey: The Benefits of Humanistic disciplines Education: An Probe of Causality and Individual Perceptions composed by Dr. Barbara Airulla, we besides see similar benefits persons receive from holding an instruction in the humanistic disciplines. She asked respondents composed of both parents and non-parents whether they believe that instruction in the ocular humanistic disciplines and music contributes to other academic success? ( Such as higher classs or trial tonss in other topics like math or scientific discipline. ) Of the 150 respondents, she received responses of 125 yes ( 93.98 % ) , 8 no ( 6 % ) , and 17 non-responses. In the 2nd portion to this inquiry she besides asked them to bespeak one ground why they believe ocular humanistic disciplines and music instruction contribute to academic success ( Airulla, B. 2004 ) . The undermentioned diagram indicates the responses she received: Figure 1: Reasons why you believe instruction in ocular humanistic disciplines and music contributes to other academic success. ( Airulla, B. , 2004 ) . Many of these responses are similar to the results found through research and you can see that people genuinely believe a music instruction has a great impact on persons. However, on the contrast to these old surveies and many more, some school principals, overseers, school board functionaries, and instructors are still blinded by the valuable effects of music to a pupil ‘s instruction.Benefits of a Music EducationIn bend, what benefits are gained through music and how can humans win from analyzing it? Thinking abilities grow, pupils are more occupied in what they study, and they learn from each other through music. The development of certain parts of the encephalon along with IQ, reading degree, and a sharper memory come with exposure to music. Every clip instrumentalists perform, they have to modify the round, pacing, phrases, pitch, temper, and musical technique. From these uninterrupted challenges, the encephalon becomes improbably skilled at multitasking ( Children ‘ s Music Workshop, 2012 ) . When the Princeton, New Jersey based College Entrance Examination Board looked at SAT tonss of college-bound high school seniors, they discovered instrumentalists of 6 old ages or more scored 57 points higher on the trials verbal subdivision and 41 points higher in math than their non-art equals ( Children ‘s Music Workshop, 2012 ) . A survey shown by the Glen Burnie County School District in Maryland stated test tonss of 223 4th through 6th class pupils both involved in and missing music instruction showed no distinguishable difference. Those pupils involved within the music plan showed no marks of increased trial tonss than their non-music equals ( Corral, 1998 ) . However, this survey was conducted on kids between 4th through 6th grade pupils who have non had adequate musical experience in order to efficaciously heighten trial tonss ; hence, this information is irrelevant and can non be used to find the consequence of music on its participants. In spring 2006, the Harris Interactive canvass of high school principals noted a drastically higher category attending of 93.3 per centum and a 90.2 per centum graduation rate among schools with music plans. Unlike their antagonist, schools without music classs merely obtain attending of 84.9 per centum among pupils and a graduation rate of 72.9 per centum ( NAfME, 2012 ) . These are simply a few of the legion successes of a music instruction. Figure 2: Comparison of graduation and attending rates among schools with a music instruction vs. those without a music instruction. Not merely does music instruction make pupils smarter, it besides teaches subject and heed. Learning to play an instrument can be compared to larning another linguistic communication and can be disputing at times. Students have to be disciplined in order to get the hang playing an instrument ; it takes clip and committedness to put out clip each twenty-four hours to pattern, pattern, and pattern some more ( Petress, 2005 ) . Musicians learn workmanship as they study how inside informations are put together fastidiously and what constitutes good, as opposed to mediocre work. The National Education Longitundinal Study showed these criterions of excellence have had an consequence on the enormous sums of academic and honor awards pupils receive ( Children ‘s Music Workshop, 2012 ) . School systems ought to demand music instruction as a new degree of excellence and a rank of virtue all parents and schools should desire for kids.How Music Education Helps StudentsAlong with the old be nefits of a music instruction, music besides allows pupils to show feelings and have a self-identity. Band and orchestra members are less likely to smoke, imbibe intoxicant, or experiment with other drugs than other non-music childs ( Frisco School of Music, 2012 ) . A sense of accomplishment comes with music. Once pupils achieve their end, the satisfaction they feel is invaluable and is surely something to be proud of. Music helps us happen our self-identity and allows us to see into our privy inner universes by showing which type ( s ) of music we enjoy. For this ground, group individualism through music is both wide and limited. An instruction in music may let a pupil to be included in multiple coteries, every bit good as excepting them. The bonding of members in a group who portion common thoughts, beliefs, and behaviours is formed through this construct. Music besides helps insulate and divide one coterie from another. Hazard taking through music public presentation teaches pup ils to get the better of fright and prevent it from going a job subsequently in life ( Children ‘s Music Workshop, 2012 ) . Learning through music allows pupils to be themselves in their ain manner. Should n't every parent want this for their kid? Have you of all time wondered how a kid sees the universe? A music instruction offers kids the ability to detect the universe otherwise. Cultural instruction is continually learned and encouraged through exhibition of the humanistic disciplines and humanistic disciplines ( NAfME, 2012 ) . Students who study music learn to esteem and hold understanding towards people of other point of views, civilizations, and imposts. Analyzing music is like larning a 2nd linguistic communication in which most people speak. Many instrumentalists have taken a piece of music and translated it into several linguistic communications and manners. It ‘s through thoughts like these, that people all around the universe are introduced to something they usually would ne'er take part in or listen to. Music cuts across racial, cultural, societal, educational, and economical limitations ( Petress, 2005 ) . It is of import that in every instruction system, schools teach pupils about different ways of life, a nd non merely the one pupils live in.Music Teaches Lifelong SkillsNot merely does a music instruction aid pupils throughout school, it teaches accomplishments which are indispensable in the work force and community. Increasingly obliging techniques established through an instruction in the humanistic disciplines are appealing to the direction and aggressiveness of the state ‘s work force ( Children ‘s Music Workshop, 2012 ) . Common art sections of the originative work force are multiplying at a rate of more than double the remainder of the state ‘s work force. Such units include painters, histrions, lensmans and writers. Within these sections are persons employed in architecture, manner design, movie, picture, music, and package development ( NAfME, 2012 ) . The humanistic disciplines besides help excite economic activity by pulling tourers and increasing the community ‘s local revenue enhancement base ( Children ‘s Music Workshop, 2012 ) . These are s ome of the many advantages music instruction offers pupils in the work force and throughout their life-time. Students ‘ point of views on how studying music helps them, is an of import factor in schools taking to necessitate music in the course of study. Members of New Milford ‘s music plan and Wagner ‘s orchestra Pitt say their music accomplishments have helped them in faculty members such as math and scientific discipline. Based on neurological trials, Mr. Michael Pitt, president of the music section at Robert Wagner Middle School and president of the Music Education Association of New York City, says â€Å" it has been proven that pupils who have a hands-on experience in the humanistic disciplines achieve more elaborate job work outing in faculty members than pupils who are introduced to computing machines at immature age † ( Gill et al. , 2000 ) . By an extra and advanced vocabulary and a more focussed schoolroom, pupils have besides learned to hold forbearance in what they do ( Rubin, 2007 ) . Obviously, pupils enjoy music and like that it ‘s something th ey can make recreationally, but music instruction is besides a category to look frontward to at school. It is different from the needed categories because there are no trials and few dull minutes where pupils sit at that place listening to the instructor. Students need this chance to take a interruption from the required nucleus categories in which most of them dread. New Milford ‘s high school pupils involved in their music instruction plan said music is what defines them. Music helps them show themselves and demo others music is cool and non simply for geeks ( Rubin, 2007 ) . A music instruction really does what is said to be true about it ; trial tonss have improved and are go oning to come on, while music helps pupils find their self-image. An instruction in music besides relieves emphasis on pupils, and broadens their skylines. Former President Bill Clinton provinces, â€Å" music is about communicating, creativeness, and cooperation, and by analyzing music in school, pupils have the chance to construct on these accomplishments, enrich their lives, and see the universe from a new position. † This quotation mark sums up the cardinal points of why an instruction in music is good to most people. We must proclaim and promote a new attack for other pupils, parents, and schools find a manner to appreciate music before it is excessively late ( Children ‘s Music Workshop, 2012 ) .DecisionAn of import decision to come out of this treatment is we as a society return music for granted. Human existences need to be educated with all types of instruction in order to accomplish maximal human potency. By back uping this position, you could assist put music in the nucleus of the school course of study. All pupils ought to hold an chance to meet and spread out their endowments in all educational systems. Schools demand to maintain and implement the humanistic disciplines in instruction because they set up in pupils a life style that last a life-time.RecommendationsIn order to promote music instruction in schools, I recommend that people do their best to inform the populace and school decision makers about the impacts a music instruction has on persons. The more advocators we have to assist back up the position of set uping music instruction in the school curriculum the better opportunity persons will hold at wining in life. By prolonging this position we can greatly impact the lives of future coevalss for old ages to come.

Thursday, November 7, 2019

Comparing of the Rate of Injuries and Illnesses in the Three Facilities of Eyeglass Manufacturing

Comparing of the Rate of Injuries and Illnesses in the Three Facilities of Eyeglass Manufacturing Introduction The present case focuses on evaluation of potential ergonomic hazards at 3 eyeglass manufacturing plants. Managers of the company addressed researchers and asked them to analyze potential hazards in surfacing and finishing departments of the three facilities. The researchers evaluated risk factors for developing Work-related Musculoskeletal Disorder (WMSD) in employees. Since the managers were interested in finishing and surfacing departments, the researchers paid special attention to surfacing and finishing tasks.Advertising We will write a custom essay sample on Comparing of the Rate of Injuries and Illnesses in the Three Facilities of Eyeglass Manufacturing specifically for you for only $16.05 $11/page Learn More According to the report, there are a lot of risk factors and hazards which have to be addressed (Ramsey Tapp, 2012). The researchers also stress that some employees have health problems including WMSDs. The researchers also compare the rate of injuries and illnesses in the three facilities and the rest of eyeglass manufacturing plants. Notably, the researchers provide specific recommendations which can help address the issue. The purpose of the present paper is to evaluate the report in question. Methods The researchers used a number of evaluation criteria to assess the hazards. First of all, the researchers focused on WMSDs symptoms. The researchers obtained the data through observations and videotaping. As has been mentioned above, the researchers concentrated on finishing and surfacing tasks. Apart from this, the researchers interviewed employees personally. The researchers focused on employees’ knowledge of ergonomics and safety measures as well as employees’ physical states and health-related complaints. The researchers also analyzed medical records of employees. Finally, the researchers analyzed records concerning injuries and accidents at the three facilities. Results The researchers repor t that employees are exposed to a number of hazards at the department in question. At this point, it is necessary to note that the researchers claim that the rate of illnesses at the facilities has decreased recently. Nonetheless, the rate is very high. The most common hazards are associated with â€Å"awkward postures, forceful exertions, and repetitive motion tasks†, twisting and bending (Ramsey Tapp, 2012, p. 9). The researchers stress that such conditions increased employees’ risk of developing â€Å"shoulder, arm, elbow, forearm, wrist, hand, and finger WMSDs† (Ramsey Tapp, 2012, p. iii). The researchers also note that employees often have lower back disorders. Noteworthy, the researchers state that the rate of injuries as well as illnesses at the three facilities in question is higher than rates at the majority of the US eyeglass manufacturing plants. Finally, the researchers note that employees do not have sufficient knowledge on ergonomics and safety m easures at the three plants.Advertising Looking for essay on business economics? Let's see if we can help you! Get your first paper with 15% OFF Learn More Recommendations The researchers provide a set of recommendations for managers and employees. Thus, managers should have a working height changed to 27-62 inches. This will diminish risk factors of developing back and shoulder WMSDs (Ramsey Tapp, 2012). According to researchers, it is also essential to provide employees with adjustable tables. This will make the working place appropriate for each employee. The researchers emphasize that it is important to â€Å"rotate† employees to different tasks (Ramsey Tapp, 2012, p. 22). This will diminish negative effects of doing repetitive tasks. Furthermore, the researchers note that managers should educate employees on WMSDs and major principles of ergonomics. Finally, the researchers stress that managers should encourage employees to report about WMSD sy mptoms or potential hazards. As far as employees are concerned, the researchers state that they also can and should improve the situation. One of the most basic recommendations made is to follow easy rules while lifting heavy objects. Besides, employees should learn more about the use of adjustable equipment and participate in ergonomic committees. Finally, employees should report about injuries or potential hazards to managers. Discussion It is necessary to note that recommendations made and points discuss are covered in other researches. For instance, the researchers in the NIOSH emphasize that educating employees on principles of ergonomics is crucial. Likewise, McCauley-Bush (2011) states that it is essential for managers as well as employees to learn more about ergonomics as this will potentially reduce risk factors at workplace. Apart from education and training, there is another important concern in the field. Thus, diversity at workplace is one of the most burning issues in the contemporary business world. The researchers touched upon the issue in their report. Thus, they claim that adjustable equipment can significantly improve the situation. Hussain et al. (2012) note that it is not enough to adjust equipment to fit employees’ different physical features. The researchers claim that it is crucial to take into account such aspects as age, culture, gender, skills, background, marital status, etc. The researchers state that all these aspects should be considered while developing ergonomic plans (Hussain et al., 2012). Admittedly, this can reduce risk factors at manufacturing facilities. It is necessary to add that researchers of the NIOSH did not consider technologic advances in ergonomics. At present, researchers make use of certain software to develop effective ergonomic plans. For example, Qutubuddin, Hebbal and Kumar (2012) state that user-friendly applications can help managers develop efficient ergonomic plans for various settings. Admittedl y, software can be used to analyze various hazards and develop effective strategies to minimalize potential risk factors.Advertising We will write a custom essay sample on Comparing of the Rate of Injuries and Illnesses in the Three Facilities of Eyeglass Manufacturing specifically for you for only $16.05 $11/page Learn More Conclusion On balance, it is possible to note that the report in question is effective. First of all, it is necessary to state that the report is based on comprehensive research. The researchers analyzed substantial amount of data. Remarkably, the researchers used different methods to obtain data. This makes the research comprehensive and plausible. It identifies major hazards and risk factors. The report also contains a set of recommendations aimed at improving the situation. Notably, the report includes recommendations for both managers as well as employees. The recommendations provided are clear and concise. They are easy to implemen t. It is also necessary to note that the report is clear and concise. The data obtained are presented in tables. This makes the report clear and detailed. The information is arranged in a logical way and the major points are highlighted separately. The report can have a wide audience. Top management, supervisors, employees, educators and students can benefit from the present report. As far as I am concerned, the report is comprehensive and detailed. Though, I would add certain information on software which can be used to develop an ergonomic plan. I would also expand the section on training. People are still unaware of effective ergonomic training. Hence, managers may need assistance in developing efficient training. It is also unclear what employees could discuss at ergonomic committees. Irrespective of the missing information, the report is still effective and can help the company improve the situation. I have learnt a lot about analyzing particular workplaces in terms of safety a nd ergonomics. I have also acknowledged that managers start paying the necessary attention to ergonomics and employees’ safety. Reference List Hussain, A.H., Marshall, R., Summerskill, S., Case, K. (2012). Workforce diversity and ergonomic challenges for sustainable manufacturing organizations. In S. Trzcielinski W. Karwowski (Eds.), Advances in ergonomics in manufacturing (pp. 23-33). Boca Raton, FL: CRC Press. McCauley-Bush, P. (2011). Ergonomics: Foundational principles, applications, and technologies. Boca Raton, FL: CRC Press.Advertising Looking for essay on business economics? Let's see if we can help you! Get your first paper with 15% OFF Learn More Ramsey, J.G., Tapp, L. (2012). Ergonomic evaluation of surfacing and finishing tasks during eyeglass manufacturing – Minnesota. Retrieved from https://www.cdc.gov/niosh/hhe/reports/pdfs/2010-0114-3168.pdf Qutubuddin, S.M., Hebbal, S.S., Kumar, A.C.S. (2011).Computer assisted system for enhancing the application of ergonomics in manufacturing systems. International Journal of Ergonomics, 2(1), 1-56.