doc_id
stringlengths 64
64
⌀ | url
stringlengths 38
44
⌀ | title
stringlengths 1
211
| language
stringclasses 1
value | text
stringlengths 2
383k
|
---|---|---|---|---|
null | null | Sousaphone | eng_Latn | The sousaphone () is a brass instrument in the same family as the more widely known tuba. Created around 1893 by J.W. Pepper at the direction of American bandleader John Philip Sousa (after whom the instrument was then named), it was designed to be easier to play than the concert tuba while standing or marching, as well as to carry the sound of the instrument above the heads of the band. Like the tuba, sound is produced by moving air past the lips, causing them to vibrate or "buzz" into a large cupped mouthpiece. Unlike the tuba, the instrument is bent in a circle to fit around the body of the musician; it ends in a large, flaring bell that is pointed forward, projecting the sound ahead of the player. Because of the ease of carrying and the direction of sound, it is widely employed in marching bands, as well as various other musical genres. Sousaphones were originally made out of brass but in the mid-20th century started to be made from lighter materials like fiberglass; today both types are in wide use.
History
The first sousaphone was built by James Welsh Pepper in 1893 at the request of John Philip Sousa, who was dissatisfied with the hélicons in use by the United States Marine Band. Some sources credit C.G. Conn with its construction, because of the first sousaphone he built later in 1898. Sousa wanted a tuba-like instrument that would send sound upward and over the band, much like a concert (upright) tuba. The new instrument had an oversized bell pointing straight up, rather than the directional bell of a normal hélicon.
The sousaphone was initially developed as a concert instrument rather than for marching. Sousa wanted the new instrument for the professional band which he started after leaving the Marines, and this band marched only once. Sousa mainly used sousaphones built by C.G. Conn. Although less balanced on a player's body than a helicon, because of the large spectacular bell high in the air, the sousaphone retained the tuba-like sound by widening the bore and throat of the instrument significantly. Its upright bell led to the instrument being dubbed a "rain-catcher". Some versions of this design allowed the bell to also rotate forward, projecting the sound to the front of the band. This bell configuration remained the standard for several decades and is the standard today.
The instrument proved practical for marching, and by 1908 the United States Marine Band adopted it.
Versions with the characteristic extra 90° bend making a forward-facing bell were developed in the early 1900s. Early sousaphones had bells, with bells popular in the 1920s. From the mid-1930s onward, sousaphone bells have been standardized at a diameter of . Some larger sousaphones (Monster, Grand, Jumbo, Giant or Grand Jumbo, depending on brand) were produced in limited quantities.
Construction
The sousaphone is a valved brass instrument with the same tube length and musical range as other tubas. The sousaphone's shape is such that the bell is above the tubist's head and projecting forward. The valves are situated directly in front of the musician slightly above the waist and all of the weight rests on the left shoulder. The bell is normally detachable from the instrument body to facilitate transportation and storage. Except for the instrument's general shape and appearance, the sousaphone is technically similar to a tuba.
For simplicity and light weight, modern sousaphones almost always use three non-compensating piston valves in their construction, in direct contrast to their concert counterparts' large variation in number, type, and orientation. Both the tuba and sousaphone are semi-conical brass instruments. No valved brass instrument can be entirely conical, since the middle section containing the valves must be cylindrical. While the degree of bore conicity does affect the timbre of the instrument, much as in a cornet and trumpet, or a euphonium and a trombone, the bore profile of a sousaphone is similar to that of most tubas.
To facilitate making the mouthpiece accessible to players of different height or body shapes, most sousaphones contain a detachable tubing gooseneck which arises from the lead pipe on the upwind side of the valves. One or two slightly-angled bit(s) (short tubing lengths) are inserted into the gooseneck, and then the mouthpiece is inserted into the terminal bit. This arrangement may be adjusted in height and yaw angle to place the mouthpiece comfortably at the player's lips.
Materials
Most sousaphones are manufactured from sheet brass, usually yellow or silver, with silver, lacquer, and gold plating options, much like many brass instruments. However, the sousaphone (uniquely) is also commonly seen manufactured from fiberglass, due to its lower cost, greater durability, and significantly lighter weight.
The weight of a sousaphone can be between and .
Pitch
Most modern sousaphones are made in the key of BB (Low B Flat) and like tubas (which are commonly made in pitches of BB, CC, EE, and F) the instrument's part is written in "concert pitch", not transposed by key for a specific instrument. Although sousaphones may have a more restricted range than their concert tuba counterpart (most sousaphones have 3 valves instead of 4 to reduce weight), generally they can all play the same music and usually have parts written in the bass clef and the indicated octave is played (unlike double bass or electric bass that sound an octave lower than the indicated note). Many older sousaphones were pitched in the key of E, but current production of sousaphones in that key is limited.
Varieties
Although most major instrument manufacturers have made, and many continue to make, sousaphones, Conn and King (H.N. White) instruments are generally agreed among players to be the standards against which other sousaphones are judged for tone quality and playability. Perhaps the most highly regarded sousaphone ever built is the Conn model 20K, introduced in the mid-1930s and still in production. Some players, especially those who find the 20K too heavy for marching, prefer the slightly smaller King model 1250, first made in the late 1920s and also still in production as model 2350. Historically, Holton, York and Martin sousaphones have also been considered fine horns.
Very large bore (>= 0.750 inch) sousaphones, with oversized bells as large as in diameter, were made by Conn ("Grand Jumbo" [46K (3-valve) & 48K (4-valve)]) and King ("Jumbo" [1265 (3- & 4-valve versions)] & "Giant" [1270 (3-valve) & 1271 (4-valve)]) in the mid-1920s and 1930s, and by Martin, York, & Buescher, but they disappeared from the catalogs during the Depression or at the onset of World War II. Because of their weight and cost, few were made and even fewer survive, especially the 4-valve models.
Fiberglass
In recent years, sousaphones have been available made of fiberglass reinforced plastics instead of brass. The fiberglass versions are used mainly for marching, with brass instruments being used for all other situations. Fiberglass sousaphones can be found commonly in younger marching bands, such as middle schools, due to their lightened weight load. Depending on the model, the fiberglass version does not have as dark and rich a tone as the brass (King fiberglass sousaphones tended to have smooth fiberglass and a tone somewhat more like a brass sousaphone; Conn fiberglass sousaphones often had rough fiberglass exteriors and a thinner sound; the Conn is also lighter).
Additional valves
In the 1920s and 1930s, four-valved sousaphones were often used by professional players, especially E sousaphones; today, however, four-valved B sousaphones are uncommon and are prized by collectors, especially those made by Conn, King (H.N. White), and Holton. Jupiter Company started production of four-valve BB sousaphones in the late 2000s, and Dynasty USA makes a four-valve BB sousaphone as well. Criticisms of the fourth valve on a sousaphone center around additional weight, although the fourth valve improves intonation and facilitates playing of the lower register.
Due to the large size of most sousaphones, the sub-contra register (for which the fourth valve is largely intended) is already covered by alternate resonances, known as "false tones" (see Tuba article). Many beginners are not aware of the false-tone resonances on their sousaphones because these notes reside in the sub-contra register, which is nearly impossible for most beginners to access. Some professionals develop a "raised embouchure" to securely play these notes. This is where either the upper or lower lip (depending on the player) takes up most of the mouthpiece area. The embouchure provides almost twice the room for vibration of the single lip (compared to the 50–50 embouchure).
Non-American sousaphones
Asian sousaphones made in China and India are gaining popularity in the street band market. In Switzerland and Southern Germany, "Guggenmusik" bands often use these instruments that provide great display and passable tone. Most are tuned in E. Brands like Zweiss with older British designs make affordable sousaphones that have broken the €500 barrier. These are mostly in the medium-bell size of . Chinese brands are mostly reverse-engineered models and quite passable.
Special effects
In large marching bands of the United States, the bell is often covered with a tight fitting cloth, called a sock, which enables the sousaphone section to spell out the school's name, initials, or mascot. The Leland Stanford Junior University Marching Band Tööbz! have a tradition of painting the front surface of their sousaphone bells with a variety of images.
Sousaphone players are also known to perform the 'flaming tubas' in which flash paper is ignited in the bell, thus making it appear as if the musician is breathing fire. David Silverman (AKA Tubatron ) developed a propane powered flaming sousaphone with a trigger valve to control an array of flame jets across the top of the bell of his horn.
The Yale Precision Marching Band has made a tradition of setting fire to the tops of the bells of their sousaphones, including in the fall of 1992 when sousaphones served as the "candles" of a "wedding cake" formed by the band when two band alumni were married during a halftime show. They also utilize what they refer to as the "Überphone", a sousaphone that was disassembled from its coiled format and welded back together on a twelve-foot frame to extend straight up from the player's shoulders.
College marching bands
John Philip Sousa was a benefactor of the University of Illinois music program and a friend of the university's Director of Bands Albert Austin Harding. The Marching Illini became the first band to march in a football halftime show, and were the first band to use sousaphones on the field.
The sousaphone sections of some marching bands have developed specialized performance traditions. The University of California Marching Band Bass section traditionally "struts" during the band's pregame show. During the "strut" the section separates from the rest of the band, circles the North goal post, and rejoins the band to complete the Script Cal. The University of Southern California Trojan Marching Band sousaphones play John Williams' "Imperial March" from Star Wars in single file when crossing streets on their way to and from performances on the USC campus. When The Ohio State University Marching Band performs its traditional Script Ohio formation, a senior sousaphone player dots the "i".
The Fightin' Texas Aggie Band sousaphone section (called "Bass Horns" within the university but never, ever "tubas") execute a distinct two-step and four-step counter-march during marching performances. During halftime performances this is accompanied (specifically for the last rank consisting all of 12 bass horns) by a "huh! huh!" from the crowd.
The University of Delaware Fightin' Blue Hen Marching Band has several traditions involving sousaphone players. During pre-game, they branch off from the rest of the band. From here, the sousaphone players run in a snake around the field jumping to drum line cadence. At most pre-games they act out a skit as well. At post game, "In My Life" by The Beatles is played featuring a sousaphone solo while the band sings.
After every pre-game show at Florida State University when the section (known by all the marching band members as "Flush"- short for "Royal Flush") run in a circle, with their horns tipped up parallel to the ground, around the Seminole head on the field with the head drum major in the center of the circle. This is called "Flushing the field" or "Flushing the Indian Head", derived from the section's nickname "Flush".
The Marching Virginians of Virginia Tech perform a version of the Hokie Pokie featuring the sousaphone section putting their sousaphones in, taking their sousaphones out, putting their sousaphones in, and shaking them all about – followed by an all-sousaphone kick line.
The University of Toledo Sousaphone line (also called The RMB Dosbas) march off the field in a snake line after home games and performs the song "sonic boom" when the rest of the band meets back with them.
For the last 20 years The University of Idaho Vandals Marching Band Sousaphone section all wear long skirts that were originally used by the 1948 University Women's Chorus. University of Idaho Vandals Sousaphone Section
Musical genres
The sousaphone is an important fixture of the New Orleans brass band tradition and is still used in groups such as the Dirty Dozen Brass Band by Kirk Joseph. Soul Rebels Brass Band from New Orleans features sousaphone player Edward Lee Jr.
Sinaloa, a state of Mexico, has a type of music called Banda Sinaloense, and the sousaphone is used there as a tuba.
Damon "Tuba Gooding Jr." Bryson from The Roots plays the sousaphone on Late Night with Jimmy Fallon.
Nat Mcintosh is the sousaphone player and co-founder of Youngblood Brass Band, who play a mixture of traditional New Orleans style brass band music and hip hop.
The Lemon Bucket Orkestra, a Canadian self-described "Balkan-Klezmer-Gypsy-Punk-Super-Party-Band", features a sousaphone as one of their instruments.
Red Baraat, a Brooklyn-based dhol & brass band that fuses North Indian Bhangra with hip-hop, go-go and jazz music, features John Altieri on sousaphone.
Notable players
Warren G. Harding, the 29th President of the United States, was a sousaphone player who played well enough to join the band celebrating his election.
Jeanie Schroder of the band DeVotchKa plays sousaphone on several of the band's songs.
Tuba Gooding Jr. (Damon Bryson) plays sousaphone for The Roots
See also
Cornu (horn), an ancient Roman brass instrument that curved around the player's body
Saxtuba
References
External links
Virginia Tech Music Dictionary: Sousaphone
All the Instruments: Sousaphone
Sousaphone History, by Ian Church: Sousaphone-ologist
American musical instruments
B-flat instruments
Brass instruments
John Philip Sousa
Marching band instruments
Tuba |
null | null | History of the European Union | eng_Latn | The European Union is a geo-political entity covering a large portion of the European continent. It is founded upon numerous treaties and has undergone expansions and secessions that have taken it from 6 member states to 27, a majority of the states in Europe.
Apart from the ideas of federation, confederation, or customs union such as Winston Churchill's 1946 call for a "United States of Europe", the original development of the European Union was based on a supranational foundation that would "make war unthinkable and materially impossible" and reinforce democracy amongst its members as laid out by Robert Schuman and other leaders in the Schuman Declaration (1950) and the Europe Declaration (1951). This principle was at the heart of the European Coal and Steel Community (ECSC) (1951), the Treaty of Paris (1951), and later the Treaty of Rome (1958) which established the European Economic Community (EEC) and the European Atomic Energy Community (EAEC). The ECSC expired in 2002, while the EAEC maintains a distinct legal identity despite sharing members and institutions.
The Maastricht Treaty (1992) created the European Union with its pillars system, including foreign and home affairs alongside the European Community. This in turn led to the creation of the single European currency, the euro (launched 1999). The Maastricht Treaty has been amended by the treaties of Amsterdam (1997), Nice (2001) and Lisbon (2007).
Ideas of European unity before 1945
The original development of the European Union was based on a supranational foundation that would "make war unthinkable and materially impossible"
A peaceful means of some consolidation of European territories used to be provided by dynastic unions; less common were country-level unions, such as the Polish–Lithuanian Commonwealth and Austro-Hungarian Empire.
In the Congress of Aix-la-Chapelle of 1818, Tsar Alexander, as the most advanced internationalist of the day, suggested a kind of permanent European union and even proposed the maintenance of international military forces to provide recognised states with support against changes by violence.
An example of an organisation formed to promote the association of states between the wars to promote the idea of European union is the Pan-Europa movement.
1945–1957: From ECSC to the Treaties of Rome
World War II from 1939 to 1945 saw a human and economic catastrophe that hit Europe hardest. It demonstrated the horrors of war, and also of extremism, through the Holocaust and the atomic bombings of Hiroshima and Nagasaki. Once again, there was a desire to ensure it could never happen again, particularly with the war giving the world nuclear weapons. Most European countries failed to maintain their Great Power status, with the exception of the USSR, which became a superpower after World War II and maintained the status for 45 years. This left two rival ideologically opposed superpowers.
To ensure Germany could never threaten the peace again, its heavy industry was partly dismantled (See: Allied plans for German industry after World War II) and its main coal-producing regions were detached (Saarland, Silesia), or put under international control (Ruhr area).
With statements such as Winston Churchill's 1946 call for a "United States of Europe" becoming louder, the Council of Europe was established in 1949 as the first pan-European organisation. In the year following, on 9 May 1950, the French Foreign Minister Robert Schuman proposed a community to integrate the coal and steel industries of Europe – these being the two elements necessary to make weapons of war. (See: Schuman declaration).
On the basis of that speech, France, Italy, the Benelux countries (Belgium, Netherlands and Luxembourg) together with West Germany signed the Treaty of Paris (1951) creating the European Coal and Steel Community the following year; this took over the role of the International Authority for the Ruhr and lifted some restrictions on German industrial productivity. It gave birth to the first institutions, such as the High Authority (now the European Commission) and the Common Assembly (now the European Parliament). The first presidents of those institutions were Jean Monnet and Paul-Henri Spaak respectively.
The formation of the European Coal and Steel Community was advanced by American Secretary of State George C. Marshall. His namesake plan to rebuild Europe in the wake of World War II contributed more than $100 billion in today's dollars to the Europeans, helping to feed Europeans, deliver steel to rebuild industries, provide coal to warm homes, and construct dams to help provide power. In doing so, the Marshall Plan encouraged the integration of European powers into the European Coal and Steel Community, the precursor to present-day European Union, by illustrating the effects of economic integration and the need for coordination. The potency of the Marshall Plan caused former German Chancellor Helmut Schmidt to remark in 1997 that "America should not forget that the development of the European Union is one of its greatest achievements. Without the Marshall Plan it perhaps would never have come to that."
The attempt to turn the Saar protectorate into a "European territory" was rejected by a referendum in 1955. The Saar was to have been governed by a statute supervised by a European Commissioner reporting to the Council of Ministers of the Western European Union.
After failed attempts at creating defence (European Defence Community) and political communities (European Political Community), leaders met at the Messina Conference and established the Spaak Committee which produced the Spaak report. The report was accepted at the Venice Conference (29 and 30 May 1956) where the decision was taken to organise an Intergovernmental Conference. The Intergovernmental Conference on the Common Market and Euratom focused on economic unity, leading to the Treaties of Rome being signed in 1957 which established the European Economic Community (EEC) and the European Atomic Energy Community (Euratom) among the members.
1958–1972: Three communities
The two new communities were created separately from ECSC, although they shared the same courts and the Common Assembly. The executives of the new communities were called Commissions, as opposed to the "High Authority". The EEC was headed by Walter Hallstein (Hallstein Commission) and Euratom was headed by Louis Armand (Armand Commission) and then Étienne Hirsch. Euratom would integrate sectors in nuclear energy while the EEC would develop a customs union between members.
Throughout the 1960s tensions began to show with France seeking to limit supranational power and rejecting the membership of the United Kingdom. However, in 1965 an agreement was reached to merge the three communities under a single set of institutions, and hence the Merger Treaty was signed in Brussels and came into force on 1 July 1967 creating the European Communities. Jean Rey presided over the first merged Commission (Rey Commission).
While the political progress of the Communities was hesitant in the 1960s, this was a fertile period for European legal integration. Many of the foundational legal doctrines of the Court of Justice were first established in landmark decisions during the 1960s and 1970s, above all in the Van Gend en Loos decision of 1963 that declared the "direct effect" of European law, that is to say, its enforceability before national courts by private parties. Other landmark decisions during this period included Costa v ENEL, which established the supremacy of European law over national law and the "Dairy Products" decision, which declared that general international law principles of reciprocity and retaliation were prohibited within the European Community. All three of these judgments were made after the appointment of French judge Robert Lecourt in 1962, and Lecourt appears to have become a dominant influence on the Court of Justice over the 1960s and 1970s.
1973–1993: Enlargement to Delors
After much negotiation, and following a change in the French Presidency, Denmark, Ireland and the United Kingdom (with Gibraltar) eventually joined the European Communities on 1 January 1973. This was the first of several enlargements which became a major policy area of the Union (see: Enlargement of the European Union).
In 1979, the European Parliament held its first direct elections by universal suffrage. 410 members were elected, who then elected the first female President of the European Parliament, Simone Veil.
A further enlargement took place in 1981 with Greece joining on 1 January, six years after applying. In 1982, Greenland voted to leave the Community after gaining home rule from Denmark (See also: Special member state territories and the European Union). Spain and Portugal joined (having applied in 1977) on 1 January 1986 in the third enlargement.
Recently appointed Commission President Jacques Delors (Delors Commission) presided over the adoption of the European flag by the Communities in 1986. In the first major revision of the treaties since the Merger Treaty, leaders signed the Single European Act in February 1986. The text dealt with institutional reform, including extension of community powers – in particular in regarding foreign policy. It was a major component in completing the single market and came into force on 1 July 1987.
In 1987 Turkey formally applied to join the Community and began the longest application process for any country. After the 1988 Polish strikes and the Polish Round Table Agreement, the first small signs of opening in Central Europe appeared. The opening of a border gate between Austria and Hungary at the Pan-European Picnic on August 19, 1989 then set in motion a peaceful chain reaction, at the end of which there was no longer a GDR and the Eastern Bloc had disintegrated. Otto von Habsburg and Imre Pozsgay saw the event as an opportunity to test Mikhail Gorbachev`s reaction to an opening of the Iron Curtain. In particular, it was examined whether Moscow would give the Soviet troops stationed in Hungary the command to intervene. But with the mass exodus at the Pan-European Picnic, the subsequent hesitant behavior of the Socialist Unity Party of East Germany and the non-intervention of the Soviet Union broke the dams. So the bracket of the Eastern Bloc was broken and as a result the Berlin Wall fell together with the whole Iron Curtain. Germany reunified and the door to enlargement to the former Eastern Bloc was opened (See also: Copenhagen Criteria).
With a wave of new enlargements on the way, the Maastricht Treaty was signed on 7 February 1992 which established the European Union when it came into force the following year.
1993–2004: Creation
On 1 November 1993, under the third Delors Commission, the Maastricht Treaty became effective, creating the European Union with its pillar system, including foreign and home affairs alongside the European Community. The 1994 European elections were held resulting in the Socialist group maintaining their position as the largest party in Parliament. The Council proposed Jacques Santer as Commission President but he was seen as a second choice candidate, undermining his position. Parliament narrowly approved Santer but his commission gained greater support, being approved by 416 votes to 103. Santer had to use his new powers under Maastricht to flex greater control over his choice of Commissioners. They took office on 23 January 1995.
On 30 March 1994, accession negotiations concluded with Austria, Sweden and Finland. Meanwhile, Norway, Iceland and Liechtenstein joined the European Economic Area (which entered into force on 1 January 1994), an organisation that allowed European Free Trade Association states to enter the Single European Market. The following year, the Schengen Agreement came into force between seven members, expanding to include nearly all others by the end of 1996. The 1990s also saw the further development of the euro. 1 January 1994 saw the second stage of the Economic and Monetary Union of the European Union begin with the establishment of the European Monetary Institute and at the start of 1999 the euro as a currency was launched and the European Central Bank was established. On 1 January 2002, notes and coins were put into circulation, replacing the old currencies entirely.
During the 1990s, the conflicts in the Balkans gave impetus to developing the EU's Common Foreign and Security Policy (CFSP). The EU failed to react during the beginning of the conflict, and UN peacekeepers from the Netherlands failed to prevent the Srebrenica massacre (July 1995) in Bosnia and Herzegovina, the largest mass murder in Europe since the Second World War. The North Atlantic Treaty Organization (NATO) finally had to intervene in the war, forcing the combatants to the negotiation table. The early foreign policy experience of the EU led to foreign policy being emphasised in the Treaty of Amsterdam (which created the High Representative).
However, any success was overshadowed by the budget crisis in March 1999. The Parliament refused to approve the Commission's 1996 community's budget on grounds of financial mismanagement, fraud and nepotism. With Parliament ready to throw them out, the entire Santer Commission resigned. The post-Delors mood of euroscepticism became entrenched with the Council and Parliament constantly challenging the Commission's position in coming years.
In the following elections, the Socialists lost their decades-old majority to the new People's Party and the incoming Prodi Commission was quick to establish the new European Anti-fraud Office (OLAF). Under the new powers of the Amsterdam Treaty, Prodi was described by some as the 'First Prime Minister of Europe'. On 4 June, Javier Solana was appointed Secretary General of the Council and the strengthened High Representative for the Common Foreign and Security Policy admitted the intervention in Kosovo – Solana was also seen by some as Europe's first Foreign Minister. The Nice Treaty was signed on 26 February 2001 and entered into force on 1 February 2003 which made the final preparations before the 2004 enlargement to 10 new members.
2004–2009: Consolidation and impact of the Great Recession
On 10–13 June 2004, the 25 member states participated in the largest trans-national election in history (with the second largest democratic electorate in the world). The result of the sixth Parliamentary election was a second victory for the European People's Party-European Democrats group. It also saw the lowest voter turnout of 45.5%, the second time it had fallen below 50%. On 22 July 2004, José Manuel Barroso was approved by the new Parliament as the next Commission President. However, his new team of 25 Commissioners faced a tougher road. With Parliament raising objections to a number of his candidates he was forced to withdraw his selection and try once more. The Prodi Commission had to extend their mandate to 22 November after the new line-up of commissioners was finally approved.
A proposed constitutional treaty was signed by plenipotentiaries from EU member states on 28 October 2004. The document was ratified in most member states, including two positive referendums. The referendums that were held in France and the Netherlands failed however, killing off the treaty. The European Council agreed that the constitution proposal would be abandoned, but most of its changes would be retained in an amending treaty. On 13 December 2007 the treaty was signed, containing opt-outs for the more eurosceptic members and no state-like elements. The Lisbon treaty finally came into force on 1 December 2009. It created the post of President of the European Council and significantly expanded the post of High Representative. After much debate about what kind of person should be President, the European Council agreed on a low-key personality and chose Herman Van Rompuy while foreign policy-novice Catherine Ashton became High Representative.
The 2009 elections again saw a victory for the European People's Party, despite losing the British Conservatives who formed a smaller eurosceptic grouping with other anti-federalist right wing parties. Parliament's presidency was once again divided between the People's Party and the Socialists, with Jerzy Buzek elected as the first President of the European Parliament from an ex-communist country.
Barroso was nominated by the Council for a second term and received backing from EPP who had declared him as their candidate before the elections. However, the Socialists and Greens led the opposition against him despite not agreeing on an opposing candidate. Parliament finally approved Barroso II, though once more several months behind schedule.
In 2007, the fifth enlargement completed with the accession of Romania and Bulgaria on 1 January 2007. Also, in 2007 Slovenia adopted the euro, Malta and Cyprus in 2008 and Slovakia in 2009. However trouble developed with existing members as the eurozone entered its first recession in 2008. Members cooperated and the ECB intervened to help restore economic growth and the euro was seen as a safe haven, particularly by those outside such as Iceland.
2009–2016: European crisis
However, with the risk of a default in Greece, Ireland, Portugal and other members in late 2009–10, eurozone leaders agreed to provisions for loans to member states who could not raise funds. Accusations that this was a U-turn on the EU treaties, which rule out any bail out of a euro member in order to encourage them to manage their finances better, were countered by the argument that these were loans, not grants, and that neither the EU nor other Member States assumed any liabilities for the debts of the aided countries. With Greece struggling to restore its finances, other member states also at risk and the repercussions this would have on the rest of the eurozone economy, a loan mechanism was agreed. The crisis also spurred consensus for further economic integration and a range of proposals such as a European Monetary Fund or federal treasury.
The European Union received the 2012 Nobel Peace Prize for having "contributed to the advancement of peace and reconciliation, democracy and human rights in Europe." The Nobel Committee stated that "that dreadful suffering in World War II demonstrated the need for a new Europe [...] today war between Germany and France is unthinkable. This shows how, through well-aimed efforts and by building up mutual confidence, historical enemies can become close partners." The Nobel Committee's decision was subject to considerable criticism.
On 1 July 2013, Croatia joined the EU, and on 1 January 2014 the French Indian Ocean territory of Mayotte was added as an outermost region.
2016–2020: Brexit
On 23 June 2016, the citizens of the United Kingdom voted to withdraw from the European Union in a referendum and subsequently became the first and to date only member to trigger Article 50 of the Treaty on European Union (TEU). The vote was in favour of leaving the EU by a margin of 51.9% in favour to 48.1% against. The UK's withdrawal was completed on 31 January 2020.
2020–present: Impact of the COVID-19 pandemic
After the economic crisis caused by the COVID-19 pandemic the European leaders agreed for the first time to emit common debt to finance the European Recovery Program called Next Generation EU (NGEU).
Structural evolution
See also
Timeline of European Union history
Enlargement of the European Union
History of the euro
History of the Common Security and Defence Policy
List of European Councils
List of presidents of the institutions of the European Union
Founding fathers of the European Union
History of the European Union in Brussels
History of the location of EU institutions
History of Europe
House of European History
Virtual Centre for Knowledge on Europe Multimedia source collections on EU History
Wider European history between World War II and the fall of Communism in Europe 1945–1992
End of World War II in Europe
Cold War
Revolutions of 1989
German reunification
Yugoslavia and the European Economic Community
Breakup of Yugoslavia
Wider European history after the creation of the European Union
Maastricht Treaty
History of the European Union (1993–2004)
History of the European Union (since 2004)
Eurozone
2006 Montenegrin independence referendum
2008 Kosovo declaration of independence from Serbia
European sovereign-debt crisis
References
Further reading
Anderson, P. The New Old World (Verso 2009)
Berend, Ivan T. The History of European Integration: A New Perspective. (Routledge, 2016).
Blair, Alasdair. The European Union since 1945 (Routledge, 2014).
Chaban, N. and M. Holland, eds. Communicating Europe in Times of Crisis: External Perceptions of the European Union (2014).
Dedman, Martin. The origins and development of the European Union 1945-1995: a history of European integration (Routledge, 2006).
De Vries, Catherine E. "Don't Mention the War! Second World War Remembrance and Support for European Cooperation." JCMS: Journal of Common Market Studies (2019).
Dinan, Desmond. Europe recast: a history of European Union (2nd ed. Palgrave Macmillan), 2004 excerpt.
Fimister, Alan. Robert Schuman: Neo-Scholastic Humanism and the Reunification of Europe (2008)
Heuser, Beatrice. Brexit in History: Sovereignty or a European Union? (2019) excerpt also see online review
Hobolt, Sara B. "The Brexit vote: a divided nation, a divided continent." Journal of European Public Policy (2016): 1–19.
Jorgensen, Knud Erik, et al., eds. The SAGE Handbook of European Foreign Policy (Sage, 2015).
Kaiser, Wolfram. "From state to society? The historiography of European integration." in Michelle Cini, and Angela K. Bourne, eds. Palgrave Advances in European Union Studies (Palgrave Macmillan UK, 2006). pp 190–208.
Kaiser, Wolfram, and Antonio Varsori, eds. European Union history: themes and debates (Springer, 2016).
Koops, TJ. and G. Macaj. The European Union as a Diplomatic Actor (2015).
McCormick, John. Understanding the European Union: a concise introduction (Palgrave Macmillan, 2014).
May, Alex. Britain and Europe since 1945 (1999).
Marsh, Steve, and Hans Mackenstein. The International Relations of the EU (Routledge, 2014).
Milward, Alan S. The Reconstruction of Western Europe: 1945–51 (Univ. of California Press, 1984)
Patel, Kiran Klaus, and Wolfram Kaiser. "Continuity and change in European cooperation during the twentieth century." Contemporary European History 27.2 (2018): 165–182. online
Young, John W. Britain, France, and the unity of Europe, 1945–1951 (Leicester University, 1984).
External links
History of the EU Official Europa website
CLIOH-WORLD CLIOH-WORLD: Network of Universities supported by the European Commission (LLP-Erasmus) for the researching, teaching and learning of the history of the EU, including History of EU Integration, EU-Turkey dialogue, and linking to world history.
An Outline of the Emergence of the European Union |
null | null | William Hale Thompson | eng_Latn | William Hale Thompson (May 14, 1869 – March 19, 1944) was an American politician who served as mayor of Chicago from 1915 to 1923 and again from 1927 to 1931. Known as "Big Bill", he is the most recent Republican to have served as mayor of Chicago. Historians rank him among the most unethical mayors in American history, mainly for his open alliance with Al Capone. However, others recognize the effectiveness of his political methods and publicity-oriented campaigning, acknowledging him as a "Political Chameleon" and an effective political machine. Time magazine said in 1931, "chief credit for creating 20th Century Politics Chicago Style" should go to William Thompson.
Thompson was known for his over-the-top campaigning and uncensored language that, along with his towering height and weight, earned him the nickname "Big Bill". Upon his reelection in 1919, Thompson was at the forefront of the movement for the Chicago Public Library system and education officials to censor and ban many texts and historical recollections coming from the United Kingdom.
Though Thompson was a popular figure, his popularity collapsed after his death, when two safe-deposit boxes were found in his name containing over $1.8 million ($ million today) in cash and bonds.
Early life
Thompson was born in Boston, Massachusetts to William Hale and Medora Gale Thompson, but his family moved to Chicago when he was nine days old. Despite having been born in Boston, Thompson had strong roots in Chicago. His father, Colonel William Hale Thompson Sr., was a popular businessman within Chicago and had served as colonel in the Second Illinois Guard who had come to Chicago after serving in the United States Navy during the American Civil War. His maternal grandfather, Stephen F. Gale, the first chief of the Chicago Fire Department, had played a large part in drawing up the city's corporation charter in 1837, earning him regard as a "Chicago pioneer" by some academic journalists.
Thompson attended Chicago Public Schools, and took supplementary courses at the Fessenden School and Metropolitan College.
Thompson was meant to attend Yale but instead moved to Wyoming at the age of 14, where he became a cowboy and cattle owner and traveled across Europe, taking up ranching in Texas and New Mexico later on in his life. The experiences influenced him to add Western touches into his campaign, including his sombrero, which became a symbol for his campaign. By the age of twenty-one, he had accumulated a stake of $30,000. He returned to Chicago in 1892 after his father's death to manage his estates. Shortly after returning to Chicago, Thompson joined the Illinois Athletic Club and the Sportsmen's Club of America and quickly was appointed director-general and captain of the water polo and football teams. His six-foot frame and athletic prowess earned him the nickname "Big Bill," which stuck with him throughout his career as a politician.
Early career
Thompson served on the Real Estate Board of Chicago.
In 1900, Thompson narrowly won election as an alderman on the Chicago City Council from the 2nd Ward, his home district. Two years later, he became a member of the Cook County Board of Commissioners, serving from 1902 through 1904. During this period, Thompson formed a political alliance with Frederick Lundin, a Republican city clerk who worked under William Lorimer, a U.S. Representative from Illinois who was known for corrupt election methods. The political duo, according to most citizens, worked very well together earning them the title the "Gallagher and Shean of Chicago Politics". Thompson with his outgoing and charismatic personality paired with his towering stature and gentlemanly appearance gave him an undeniable public presence, which was completed by Lundin's cunning political ideas and projects.
First mayoralty (1915–1923)
First term
In 1915, Thompson was elected as the 41st Mayor of Chicago, beating County Clerk Robert M. Sweitzer, John H. Hill, Seymour Steadman, and Charles Thompson.
Thompson was sworn-in on April 26, 1915. In his inaugural address, Thompson spoke of his ambitions for Chicago to become "the greatest in the world", but also that his acts as mayor should not be swayed by corruption. He also emphasized the importance of public safety (as enforced by the Chicago Police Department), the improvement of public transit, secure and permanently lowered gas prices, Chicago being allowed to have Home rule and more efforts being placed into Chicago's commercial interests in order to create jobs and improve the city's economy. His efforts to expand and publicly improve the streets of Chicago earned him another nickname of "Big Bill the Builder". In his time as mayor, he oversaw the completion of the Michigan Avenue link bridge, the Twelfth Street widening, and the extension and widening of Ogden Avenue. Along with his big dreams for Chicago's geographical expansion, he wished for Chicago to expand politically and economically. He believed that Chicago should be able to enforce laws on their own terms, particularly without what he claimed to be the interference of British government or totalitarian rule. He ended his inaugural address by declaring,
As Thompson entered the first term of his mayorship, he appointed Fred Lundin as chairman on the committee of patronage. Early in his mayoral career, Thompson began to amass a war chest to support an eventual run for the presidency, by charging city drivers and inspectors $3 per month.
Early in his mayoralty Thompson had to cut short a July 1915 trip to San Francisco in order to deal with the aftermath of the Eastland disaster. While Thompson was out of town, acting-mayor Moorhouse had turned the Chicago City Hall into a makeshift hospital for first aid and a morgue for bodies recovered from the tragedy. Once Thompson returned to Chicago he organized and heavily promoted a relief fund and ordered an investigation into the casual negligence responsible for the tragedy.
In 1915, a delegation of civic-oriented women, headed by Mary McDowell, urged Thompson to appoint an well-qualified woman to the city's new office of "commissioner of public welfare". Thompson did appoint a woman. However, instead of a woman qualified by a public welfare background, he appointed Louise Osborn Rowe, a Republican Party worker and loyalist. Within a year of her appointment, Rowe was charged with operating a kickback scheme in the department, and was forced to resign in 1916. This post would remain vacant until the mayoralty of Thompson's successor.
Thompson gained national attention and condemnation for his neutral attitude toward the events of World War I. By declining the visit of the French Mission to Chicago and refusing to control or act against anti-war or anti-conscription meetings, Thompson is "credited with characterizing Chicago as the sixth German city of the world," also earning the nickname "Kaiser" Bill Thompson. Thompson sought to further endear himself to the city's German and Irish populations by positioning himself as anti-British. These facts later went on to hurt his chances in his 1918 U.S. Senate campaign.
In 1916, he became a member of the Republican National Committee. He would continue to serve on the committee until 1920.
Second term
Thompson was reelected mayor in 1919, beating out Robert Sweitzer once again along with Adolph S. Carm, John Collins, John Fitzpatrick, and Maclay Hoyne. Thompson was said to have had control of the 75,000 black voters in his day. In his campaign he claimed to be an advocate for the people against public utility companies and the rich who avoided taxes. This inspired Thompson to enforce a five-cent streetcar fare to promote his campaign, which was also used to threaten the action of streetcar companies. Eventually, however, despite his protests, the fare was raised to seven cents.
In his second inaugural address on April 28, 1919, Thompson looked towards drastically expanding Chicago, saying that "Chicago is greater than some nations". This expansion included the extension and widening of streets to cross over more of the city, new post offices, freight terminals, playgrounds, bridges, and more. Also, due to the rapidly changing city, Thompson proposed a zoning bill to regulate and create commercial, industrial, and residential areas. Among the other issues he claimed he would address were telephone prices and service quality, the expansion of the Chicago Police Department, jobs for returning soldiers, lowering the cost of living, and restoring the jobs of Public School representatives who were removed by the Supreme Court.
Early into his second term, the city dealt with the Chicago race riot of 1919.
At the 1920 Republican National Convention Thompson helped to block his one-time ally Frank Lowden from capturing the nomination.
Thompson declined to run for reelection in 1923 and he was succeeded by William Emmett Dever. Thompson left office as Mayor on April 16, 1923.
Chairman of the Illinois Waterways Commission
While out of office, Thompson was appointed chairman of the Illinois Waterways Commission. He used his position to remain relevant in the media, involving himself in civic suits and campaigning for the Lakes-to-Gulf waterway project: to build a waterway from the Great Lakes to the Gulf of Mexico. Promoting both the project and himself, Thompson set off on a "scientific" expedition (to be extensively covered by the media), which he set off to the South Seas in order to find a tree-climbing fish on July 5, 1924. Attracting more attention, Thompson placed a $25,000 bet on his success, but no one participated.
Second mayoralty (1927–1931)
Thompson ran again in 1927 during a citywide gang war, aiming to unseat his successor, William Dever. Always a flamboyant campaigner, Thompson held a primary-election debate between himself and two live rats which he used to portray his opponents. Pledging to clean up Chicago and remove the crooks, Thompson instead turned his attention to the reformers, whom he considered the real criminals. According to Thompson, the biggest enemy the United States had was King George V of the United Kingdom. Thompson promised his supporters, many of whom were Irish, that if they ever met, Thompson would punch the king in the nose, or at other times, that he would arrest His Majesty. Upon his victory over Dever, Thompson's floating speakeasy, outwardly known as the Fish Fans Club, docked at Belmont Harbor. It was flooded with his supporters, so many so that the boat itself sank beneath the weight.
In his inaugural address on April 18, 1927, Thompson addressed the importance of remedying crime in Chicago, saying,
Thompson expressed his desire to remove Superintendent William McAndrew from the public schooling system, and restore what he called the "true history of George Washington" while exposing "the treason and propaganda which insidiously have been injected into our schools and other educational institutions". He also went on to enforce other issues he had addressed in previous speeches, like the issue of public transit, playgrounds, and the general upkeep and expansion of Chicago in an effort to aid property owners and increase residential income and revenue for the city as a whole. In August 1927, the Chicago Board of Education, now under Thompson's influence after he appointed a number new members, voted to charge McAndrew with insubordination and lack of patriotism, suspending him pending an administrative hearing held by the board. The administrative hearing would last months, and the Chicago Board of Education would find McAndrew guilty. The Cook County Superior Court would later void this decision.
Al Capone's support was pivotal to Thompson's return to the mayor's office. During Thompson's second term, the "Pineapple Primary" took place on April 10, 1928, so-called because of the hand grenades thrown at polling places to disrupt voting. The Pinneaple Primary saw candidates backed by Thompson face Charles Deneen in the Republican primary election. Another infamous instance of gang activity that took place during Thompson's third term was St. Valentine's Day Massacre.
Thompson blamed Ruth Hanna McCormick's lack of support for his loss at the 1928 Republican National Convention, and he returned the favor during her 1930 campaign for the United States Senate by endorsing against her in the general eletion. Thompson had had a longstanding rivalry with the McCormicks. He intensely disliked Robert R. McCormick who published the Chicago Tribune. U.S. Senator Medill McCormick was the publisher's brother, and Ruth Hanna McCormick was Medill McCormick's wife.
Amid growing discontent with Thompson's leadership, particularly in the area of cleaning up Chicago's reputation as the capital of organized crime, he was defeated in 1931 by Democrat Anton Cermak. Cermak was an immigrant from Bohemia, and Thompson used this fact to belittle him with ethnic slurs such as:
I won't take a back seat to that Bohunk, Chairmock, Chermack or whatever his name is.
Tony, Tony, where's your pushcart at?
Can you picture a World's Fair mayor with a name like that?
Cermak replied, "He doesn't like my name...It's true I didn't come over on the Mayflower, but I came over as soon as I could," which was a sentiment to which ethnic Chicagoans (especially its large Bohemian population) could relate, so Thompson's slurs largely backfired.
After Thompson's defeat, the Chicago Tribune wrote,
Thompson left office April 9, 1931.
Subsequent career
In 1936, Thompson ran for the office of Illinois governor on the "Union Progressive" ballot line against Democratic incumbent Henry Horner and Republican nominee C. Wayland Brooks. He received only three percent of the vote. In 1939, Thompson ran in the Republican primary for mayor of Chicago and was soundly defeated by a 77% to 23% margin against future Governor Dwight Green.
Personal life
In 1901, Thompson married Mary "Maysie" Walker Wyse, a secretary in his father's office. The two never had children.
Death
William Hale Thompson died on March 19, 1944, at the Blackstone Hotel at the age of 74. He was buried in Oak Woods Cemetery in a solid bronze casket.
Despite the fact that many loved Thompson and enjoyed his various political antics, few people attended his funeral, and one reporter noted that there was not "a flower nor a fern to be seen".
Upon Thompson's death, two safe deposit boxes in his name were discovered to contain nearly $1.84 million ($ million today) in cash. Once the money was uncovered, the Internal Revenue Service took their share in taxes, and Maysie Thompson lived off of the rest until her death in 1958.
References
Further reading
External links
Mayors of Chicago
Chicago City Council members
Members of the Cook County Board of Commissioners
Businesspeople from Chicago
Al Capone associates
Illinois Independents
Illinois Republicans
Politicians from Boston
1868 births
1944 deaths
American political bosses from Illinois
Republican National Committee members |
null | null | Scale (anatomy) | eng_Latn | In most biological nomenclature, a scale (; ) is a small rigid plate that grows out of an animal's skin to provide protection. In lepidopteran (butterfly and moth) species, scales are plates on the surface of the insect wing, and provide coloration. Scales are quite common and have evolved multiple times through convergent evolution, with varying structure and function.
Scales are generally classified as part of an organism's integumentary system. There are various types of scales according to shape and to class of animal.
Fish scales
Fish scales are dermally derived, specifically in the mesoderm. This fact distinguishes them from reptile scales paleontologically.
Genetically, the same genes involved in tooth and hair development in mammals are also involved in scale development.
Cosmoid scales
True cosmoid scales can only be found on the Sarcopterygians. The inner layer of the scale is made of lamellar bone. On top of this lies a layer of spongy or vascular bone and then a layer of dentine-like material called cosmine. The upper surface is keratin. The coelacanth has modified cosmoid scales that lack cosmine and are thinner than true cosmoid scales.
Ganoid scales
Ganoid scales can be found on gars (family Lepisosteidae), bichirs, and reedfishes (family Polypteridae). Ganoid scales are similar to cosmoid scales, but a layer of ganoin lies over the cosmine layer and under the enamel. Ganoin scales are diamond shaped, shiny, and hard.
Within the ganoin are guanine compounds, iridescent derivatives of guanine found in a DNA molecule. The iridescent property of these chemicals provide the ganoin its shine.
Placoid scales
Placoid scales are found on cartilaginous fish including sharks and stingrays. These scales, also called denticles, are similar in structure to teeth, and have one median spine and two lateral spines. The modern jawed fish ancestors, the jawless ostracoderms and later jawed placoderms, may have had scales with the properties of both placoid and ganoid scales.
Leptoid scales
Leptoid scales are found on higher-order bony fish. As they grow they add concentric layers. They are arranged so as to overlap in a head-to-tail direction, like roof tiles, allowing a smoother flow of water over the body and therefore reducing drag. They come in two forms:
Cycloid scales have a smooth outer edge, and are most common on fish with soft fin rays, such as salmon and carp.
Ctenoid scales have a toothed outer edge, and are usually found on fish with spiny fin rays, such as bass and crappie.
Reptilian scales
Reptile scale types include: cycloid, granular (which appear bumpy), and keeled (which have a center ridge). Scales usually vary in size, the stouter, larger scales cover parts that are often exposed to physical stress (usually the feet, tail and head), while scales are small around the joints for flexibility. Most snakes have extra broad scales on the belly, each scale covering the belly from side to side.
The scales of all reptiles have an epidermal component (what one sees on the surface), but many reptiles, such as crocodilians and turtles, have osteoderms underlying the epidermal scale. Such scales are more properly termed scutes. Snakes, tuataras and many lizards lack osteoderms. All reptilian scales have a dermal papilla underlying the epidermal part, and it is there that the osteoderms, if present, would be formed.
Avian scales
Birds' scales are found mainly on the toes and metatarsus, but may be found further up on the ankle in some birds. The scales and scutes of birds were thought to be homologous to those of reptiles, but are now agreed to have evolved independently, being degenerate feathers.
Mammalian scales
An example of a scaled mammal is the pangolin. Its scales are made of keratin and are used for protection, similar to an armadillo's armor. They have been convergently evolved, being unrelated to mammals' distant reptile-like ancestors (since therapsids lost scales), except that they use a similar gene.
On the other hand, the musky rat-kangaroo has scales on its feet and tail. The precise nature of its purported scales has not been studied in detail, but they appear to be structurally different from pangolin scales.
Anomalures also have scales on their tail undersides.
Foot pad epidermal tissues in most mammal species have been compared to the scales of other vertebrates. They are likely derived from cornification processes or stunted fur much like avian reticulae are derived from stunted feathers.
Arthropod scales
Butterflies and moths - the order Lepidoptera (Greek "scale-winged") - have membranous wings covered in delicate, powdery scales, which are modified setae. Each scale consists of a series of tiny stacked platelets of organic material, and butterflies tend to have the scales broad and flattened, while moths tend to have the scales narrower and more hair like. Scales are usually pigmented, but some types of scales are metallic, or iridescent, without pigments; because the thickness of the platelets is on the same order as the wavelength of visible light the plates lead to structural coloration and iridescence through the physical phenomenon described as thin-film optics. The most common color produced in this fashion is blue, such as in the Morpho butterflies. Other colors can be seen on the sunset moth.
See also
Scale (insect anatomy)
Armour (anatomy)
Psoriasis: a long-lasting autoimmune disease characterized by patches of thin pieces of hard skin like scale.
References
Further reading
Animal anatomy
Marine biology
Integumentary system |
null | null | Russo-Japanese War | eng_Latn | The Russo-Japanese War (; ) was fought between the Empire of Japan and the Russian Empire during 1904 and 1905 over rival imperial ambitions in Manchuria and the Korean Empire. The major theatres of military operations were the Liaodong Peninsula and Mukden in Southern Manchuria, and the seas around Korea, Japan, and the Yellow Sea.
Russia sought a warm-water port on the Pacific Ocean both for its navy and for maritime trade. Vladivostok remained ice-free and operational only during the summer; Port Arthur, a naval base in Liaodong Province leased to Russia by the Qing dynasty of China from 1897, was operational year round. Since the end of the First Sino-Japanese War in 1895, Japan feared Russian encroachment on its plans to create a sphere of influence in Korea and Manchuria. Russia had demonstrated an expansionist policy east of the Urals in Siberia and the Far East from the reign of Ivan the Terrible in the 16th century.
Seeing Russia as a rival, Japan offered to recognize Russian dominance in Manchuria in exchange for recognition of Korea being within the Japanese sphere of influence. Russia refused and demanded the establishment of a neutral buffer zone between Russia and Japan in Korea north of the 39th parallel. The Imperial Japanese Government perceived a threat to their plans for expansion into mainland Asia and chose to go to war. After negotiations broke down in 1904, the Imperial Japanese Navy (IJN) opened hostilities in a surprise attack on by attacking the Russian Eastern Fleet at Port Arthur, China.
Although Russia suffered a number of defeats, Emperor Nicholas II was convinced that Russia could win if it fought on; he chose to remain engaged in the war and wait for the outcomes of certain key naval battles. After the hope of victory was quelled, he continued the war to preserve the dignity of Russia by averting a "humiliating peace." Russia ignored Japan's willingness early on to agree to an armistice and rejected the idea of bringing the dispute to the Arbitration Court at The Hague. The war eventually concluded with the Treaty of Portsmouth (), mediated by US President Theodore Roosevelt. The complete victory of the Japanese military surprised international observers and transformed the balance of power in both East Asia and Eastern Europe, resulting in Japan's emergence as a great power and Russia's decline in prestige and influence in Eastern Europe. The loss of life without victory and the humiliating defeat for the Russian Empire contributed to growing domestic unrest which culminated in the 1905 Russian Revolution and accelerated the disintegration of the Russian autocracy. The war also marked the first victory of an Asian country against a Western power in modern times.
Historical background
Modernization of Japan
After the Meiji Restoration in 1868, the Meiji government endeavoured to assimilate Western ideas, technological advances and ways of warfare. By the late 19th century, Japan had transformed itself into a modernized industrial state. The Japanese wanted to be recognized as equal with the Western powers. The Meiji Restoration had been intended to make Japan a modernized state, not a Westernized one, and Japan was an imperialist power, looking towards overseas expansionism.
In the years 1869–73, the Seikanron ("Conquer Korea Argument") had bitterly divided the Japanese elite: one faction wanted to conquer Korea immediately, another wanted to wait until Japan was further modernized before embarking on a war to conquer Korea; significantly, no one in the Japanese elite ever accepted the idea that the Koreans had the right to be independent, with only the question of timing dividing the two factions. In much the same way that Europeans used the "backwardness" of African and Asian nations as a reason for why they had to conquer them, for the Japanese elite the "backwardness" of China and Korea was proof of the inferiority of those nations, thus giving the Japanese the "right" to conquer them.
Count Inoue Kaoru, the Foreign Minister, gave a speech in 1887 saying "What we must do is to transform our empire and our people, make the empire like the countries of Europe and our people like the peoples of Europe," going on to say that the Chinese and Koreans had essentially forfeited their right to be independent by not modernizing. Much of the pressure for an aggressive foreign policy in Japan came from below, with the advocates of "people's rights" movement calling for an elected parliament also favouring an ultra-nationalist line that took it for granted the Japanese had the "right" to annex Korea, as the "people's rights" movement was led by those who favoured invading Korea in the years 1869–73.
As part of the modernization process in Japan, Social Darwinist ideas about the "survival of the fittest" were common in Japan from the 1880s onward and many ordinary Japanese resented the heavy taxes imposed by the government to modernize Japan, demanding something tangible like an overseas colony as a reward for their sacrifices.
Furthermore, the educational system of Meiji Japan was meant to train the schoolboys to be soldiers when they grew up, and as such, Japanese schools indoctrinated their students into Bushidō ("way of the warrior"), the fierce code of the samurai. Having indoctrinated the younger generations into Bushidō, the Meiji elite found themselves faced with a people who clamored for war, and regarded diplomacy as a weakness.
Pressure from the people
The British Japanologist Richard Story wrote that the biggest misconception about Japan in the West was that the Japanese people were the "docile" instruments of the elite, when in fact much of the pressure for Japan's wars from 1894 to 1941 came from the ordinary people, who demanded a "tough" foreign policy, and tended to engage in riots and assassination when foreign policy was perceived to be pusillanimous.
Though the Meiji oligarchy refused to allow democracy, they did seek to appropriate some of the demands of the "people's rights" movement by allowing an elected Diet in 1890 (with limited powers and an equally limited franchise) and by pursuing an aggressive foreign policy towards Korea.
In 1884, Japan had encouraged a coup in the Kingdom of Korea by a pro-Japanese reformist faction, which led to the conservative government calling upon China for help, leading to a clash between Chinese and Japanese soldiers in Seoul. At the time, Tokyo did not feel ready to risk a war with China, and the crisis was ended by the Convention of Tientsin, which left Korea more strongly in the Chinese sphere of influence, though it did give the Japanese the right to intervene in Korea. All through the 1880s and early 1890s, the government in Tokyo was regularly criticized for not being aggressive enough in Korea, leading Japanese historian Masao Maruyama to write: Just as Japan was subject to pressure from the Great Powers, so she would apply pressure to still weaker countries—a clear case of the transfer psychology. In this regard it is significant that ever since the Meiji period demands for a tough foreign policy have come from the common people, that is, from those who are at the receiving end of oppression at home.
Russian Eastern expansion
Tsarist Russia, as a major imperial power, had ambitions in the East. By the 1890s it had extended its realm across Central Asia to Afghanistan, absorbing local states in the process. The Russian Empire stretched from Poland in the west to the Kamchatka Peninsula in the east. With its construction of the Trans-Siberian Railway to the port of Vladivostok, Russia hoped to further consolidate its influence and presence in the region. In the Tsushima incident of 1861 Russia had directly assaulted Japanese territory.
Sino-Japanese War (1894–95)
The first major war the Empire of Japan fought following the Meiji Restoration was against China, from 1894-1895. The war revolved around the issue of control and influence over Korea under the rule of the Joseon dynasty. From the 1880s onward, there had been vigorous competition for influence in Korea between China and Japan. The Korean court was prone to factionalism, and at the time was badly divided between a reformist camp that was pro-Japanese and a more conservative faction that was pro-Chinese. In 1884, a pro-Japanese coup attempt was put down by Chinese troops, and a "residency" under General Yuan Shikai was established in Seoul. A peasant rebellion led by the Tonghak religious movement led to a request by the Korean government for the Qing dynasty to send in troops to stabilize the country. The Empire of Japan responded by sending their own force to Korea to crush the Tonghak and installed a puppet government in Seoul. China objected and war ensued. Hostilities proved brief, with Japanese ground troops routing Chinese forces on the Liaodong Peninsula and nearly destroying the Chinese Beiyang Fleet in the Battle of the Yalu River. Japan and China signed the Treaty of Shimonoseki, which ceded the Liaodong Peninsula and the island of Taiwan to Japan. After the peace treaty, Russia, Germany, and France forced Japan to withdraw from the Liaodong Peninsula. The leaders of Japan did not feel that they possessed the strength to resist the combined might of Russia, Germany and France, and so gave in to the ultimatum. At the same time, the Japanese did not abandon their attempts to force Korea into the Japanese sphere of influence. On 8 October 1895, Queen Min of Korea, the leader of the anti-Japanese and pro-Chinese faction at the Korean court was murdered by Japanese agents within the halls of the Gyeongbokgung palace, an act that backfired badly as it turned Korean public opinion against Japan. In early 1896, King Gojong of Korea fled to the Russian legation in Seoul, believing that his life was in danger from Japanese agents, and Russian influence in Korea started to predominate. In the aftermath of the flight of the king, a popular uprising overthrew the pro-Japanese government and several cabinet ministers were lynched on the streets.
In 1897, Russia occupied the Liaodong Peninsula, built the Port Arthur fortress, and based the Russian Pacific Fleet in the port. Russia's acquisition of Port Arthur was primarily an anti-British move to counter the British occupation of Wei-hai-Wei, but in Japan, this was perceived as an anti-Japanese move. Germany occupied Jiaozhou Bay, built the Tsingtao fortress, and based the German East Asia Squadron in this port. Between 1897 and 1903, the Russians built the Chinese Eastern Railway (CER) in Manchuria. The Chinese Eastern Railroad was owned jointly by the Russian and Chinese governments, but the company's management was entirely Russian, the line was built to the Russian gauge and Russian troops were stationed in Manchuria to protect rail traffic on the CER from bandit attacks. The headquarters of the CER company was located in the new Russian-built city of Harbin, the "Moscow of the Orient". From 1897 onwards, Manchuria—while still nominally part of the "Great Qing Empire"—started to resemble more and more a Russian province.
Russian encroachment
In December 1897, a Russian fleet appeared off Port Arthur. After three months, in 1898, China and Russia negotiated a convention by which China leased (to Russia) Port Arthur, Talienwan and the surrounding waters. The two parties further agreed that the convention could be extended by mutual agreement. The Russians clearly expected such an extension, for they lost no time in occupying the territory and in fortifying Port Arthur, their sole warm-water port on the Pacific coast and of great strategic value. A year later, to consolidate their position, the Russians began to build a new railway from Harbin through Mukden to Port Arthur, the South Manchurian Railroad. The development of the railway became a contributory factor to the Boxer Rebellion, when Boxer forces burned the railway stations.
The Russians also began to make inroads into Korea. By 1898 they had acquired mining and forestry concessions near the Yalu and Tumen rivers, causing the Japanese much anxiety. Japan decided to attack before the Russians completed the Trans-Siberian Railway.
Boxer Rebellion
The Russians and the Japanese both contributed troops to the Eight-Nation Alliance sent in 1900 to quell the Boxer Rebellion and to relieve the international legations besieged in the Chinese capital, Beijing. Russia had already sent 177,000 soldiers to Manchuria, nominally to protect its railways under construction. Though the Qing imperial army and the Boxer rebels united to fight against the invasion, they were quickly overrun and ejected from Manchuria. After the Boxer Rebellion, 100,000 Russian soldiers were stationed in Manchuria. The Russian troops settled in and despite assurances they would vacate the area after the crisis, by 1903 the Russians had not established a timetable for withdrawal and had actually strengthened their position in Manchuria.
Pre-war negotiations
The Japanese statesman Itō Hirobumi started to negotiate with the Russians. He regarded Japan as too weak to evict the Russians militarily, so he proposed giving Russia control over Manchuria in exchange for Japanese control of northern Korea. Of the five Genrō (elder statesmen) who made up the Meiji oligarchy, Itō Hirobumi and Count Inoue Kaoru opposed the idea of war against Russia on financial grounds, while Katsura Tarō, Komura Jutarō and Field Marshal Yamagata Aritomo favored war. Meanwhile, Japan and Britain had signed the Anglo-Japanese Alliance in 1902 – the British seeking to restrict naval competition by keeping the Russian Pacific seaports of Vladivostok and Port Arthur from their full use. Japan's alliance with the British meant, in part, that if any nation allied itself with Russia during any war against Japan, then Britain would enter the war on Japan's side. Russia could no longer count on receiving help from either Germany or France without the danger of British involvement in the war. With such an alliance, Japan felt free to commence hostilities if necessary.
The 1890s and 1900s marked the height of the "Yellow Peril" propaganda by the German government, and the German Emperor Wilhelm II () often wrote letters to his cousin Emperor Nicholas II of Russia, praising him as the "saviour of the white race" and urging Russia forward in Asia. From November 1894 onward, Wilhelm had been writing letters praising Nicholas as Europe's defender from the "Yellow Peril", assuring the Tsar that God Himself had "chosen" Russia to defend Europe from the alleged Asian threat. On 1 November 1902 Wilhelm wrote to Nicholas that "certain symptoms in the East seem to show that Japan is becoming a rather restless customer" and "it is evident to every unbiased mind that Korea must and will be Russian". Wilhelm ended his letter with the warning that Japan and China would soon unite against Europe, writing: "Twenty to thirty million Chinese, supported by a half dozen Japanese divisions, led by competent, intrepid Japanese officers, full of hatred for Christianity—that is a vision of the future that cannot be contemplated without concern, and it is not impossible. On the contrary, it is the realisation of the yellow peril, which I described a few years ago and I was ridiculed by the majority of people for my graphic depiction of it ... Your devoted friend and cousin, Willy, Admiral of the Atlantic".
Wilhelm aggressively encouraged Russia's ambitions in Asia because France, Russia's closest ally since 1894, was less than supportive of Russian expansionism in Asia, and it was believed in Berlin that German support of Russia might break up the Franco-Russian alliance and lead to a new German–Russian alliance.
The French had made it clear that they disapproved of Nicholas's forward policy in Asia; the French Premier Maurice Rouvier (in office: May to December 1887) publicly declaring that the Franco-Russian alliance applied only in Europe, not to Asia,
and that France would remain neutral if Japan attacked Russia. The American president Theodore Roosevelt (in office 1901–1909), who was attempting to mediate the Russian–Japanese dispute, complained that Wilhelm's "Yellow Peril" propaganda, which strongly implied that Germany might go to war against Japan in support of Russia, encouraged Russian intransigence. On 24 July 1905, in a letter to the British diplomat Cecil Spring Rice, Roosevelt wrote that Wilhelm bore partial responsibility for the war as "he has done all he could to bring it about", charging that Wilhelm's constant warnings about the "Yellow Peril" had made the Russians uninterested in compromise as Nicholas believed that Germany would intervene if Japan attacked.
The implicit promise of German support suggested by Wilhelm's "Yellow Peril" speeches and letters to Nicholas led many decision-makers in Saint Petersburg to believe that Russia's military weaknesses in the Far East (like the uncompleted Trans-Siberian railroad line) did not matter—they assumed that the Reich would come to Russia's assistance if war should come. In fact, neither Wilhelm nor his Chancellor Prince Bernhard von Bülow (in office: 1900–1909) had much interest in East Asia, and Wilhelm's letters to Nicholas praising him as Europe's saviour against the "Yellow Peril" were really meant to provoke change in the balance of power in Europe, as Wilhelm believed that any Russian entanglement with Japan would break up the Franco-Russian alliance and lead to Nicholas signing an alliance with Germany. This was especially the case as Germany had embarked upon the "Tirpitz Plan" and a policy of Weltpolitik (from 1897) meant to challenge Britain's position as the world's leading power. Since Britain was allied to Japan, then if Germany could manipulate Russia and Japan into going to war with each other, this in turn would allegedly lead to Russia turning towards Germany.
Furthermore, Wilhelm believed if a Russian–German alliance emerged, France would be compelled to join it. He also hoped that having Russia pursue an expansionist policy in Asia would distract and keep Russia out of the Balkans, thus removing the main source of tension between Russia and Germany's ally Austria-Hungary. During the war, Nicholas who took at face value Wilhelm's "Yellow Peril" speeches, placed much hope in German intervention on his side. More than once Nicholas chose to continue the war out of the belief that the Kaiser would come to his aid.
Despite previous assurances that Russia would completely withdraw from Manchuria the forces it had sent to crush the Boxer Rebellion by 8 April 1903, that day passed with no reduction in Russian forces in that region. In Japan, university students demonstrated both against Russia and against their own government for not taking any action. On 28 July 1903 Kurino Shin'ichirō, the Japanese minister in Saint Petersburg, was instructed to present his country's view opposing Russia's consolidation plans in Manchuria. On 3 August 1903 the Japanese minister handed in the following document to serve as the basis for further negotiations:
Mutual engagement to respect the independence and territorial integrity of the Chinese and Korean empires and to maintain the principle of equal opportunity for the commerce and industry of all nations in those countries.
Reciprocal recognition of Japan's preponderating interests in Korea and Russia's special interests in railway enterprises in Manchuria, and of the right of Japan to take in Korea and of Russia to take in Manchuria such measures as may be necessary for the protection of their respective interests as above defined, subject, however, to the provisions of article I of this agreement.
Reciprocal undertaking on the part of Russia and Japan not to impede development of those industrial and commercial activities respectively of Japan in Korea and of Russia in Manchuria, which are not inconsistent with the stipulations of article I of this agreement. Additional engagement on the part of Russia not to impede the eventual extension of the Korean railway into southern Manchuria so as to connect with the East China and Shan-hai-kwan–Newchwang lines.
Reciprocal engagement that in case it is found necessary to send troops by Japan to Korea, or by Russia to Manchuria, for the purpose either of protecting the interests mentioned in article II of this agreement, or of suppressing insurrection or disorder calculated to create international complications, the troops so sent are in no case to exceed the actual number required and are to be forthwith recalled as soon as their missions are accomplished.
Recognition on the part of Russia of the exclusive right of Japan to give advice and assistance in the interest of reform and good government in Korea, including necessary military assistance.
This agreement to supplant all previous arrangements between Japan and Russia respecting Korea.
On 3 October 1903 the Russian minister to Japan, Roman Rosen, presented to the Japanese government the Russian counter proposal as the basis of negotiations, as follows:
Mutual engagement to respect the independence and territorial integrity of the Korean Empire.
Recognition by Russia of Japan's preponderating interests in Korea and of the right of Japan to give advice and assistance to Korea tending to improve the civil administration of the empire without infringing the stipulations of article I.
Engagement on the part of Russia not to impede the commercial and industrial undertakings of Japan in Korea, nor to oppose any measures taken for the purpose of protecting them so long as such measures do not infringe the stipulations of article I.
Recognition of the right of Japan to send for the same purpose troops to Korea, with the knowledge of Russia, but their number not to exceed that actually required, and with the engagement on the part of Japan to recall such troops as soon as their mission is accomplished.
Mutual engagement not to use any part of the territory of Korea for strategical purposes nor to undertake on the coasts of Korea any military works capable of menacing the freedom of navigation in the Straits of Korea.
Mutual engagement to consider that part of the territory of Korea lying to the north of the 39th parallel as a neutral zone into which neither of the contracting parties shall introduce troops.
Recognition by Japan of Manchuria and its littoral as in all respects outside her sphere of interest.
This agreement to supplant all previous agreements between Russia and Japan respecting Korea.
During the Russian–Japanese talks, the Japanese historian Hirono Yoshihiko noted, "once negotiations commenced between Japan and Russia, Russia scaled back its demands and claims regarding Korea bit by bit, making a series of concessions that Japan regarded as serious compromises on Russia's part". The war might not have broken out had not the issues of Korea and Manchuria become linked. The Korean and Manchurian issues had become linked as the Prime Minister of Japan, Katsura Tarō (in office 1901–1906), decided if war did come, that Japan was more likely to have the support of the United States and Great Britain if the war could be presented as a struggle for free trade against the highly protectionist Russian empire, in which case, Manchuria, which was the larger market than Korea, was more likely to engage Anglo-American sympathies. Throughout the war, Japanese propaganda presented the recurring theme of Japan as a "civilized" power (that supported free trade and would implicitly allow foreign businesses into the resource-rich region of Manchuria) vs. Russia the "uncivilized" power (that was protectionist and wanted to keep the riches of Manchuria all to itself).
Emperor Gojong of Korea (King from 1864 to 1897, Emperor from 1897 to 1907) came to believe that the issue dividing Japan and Russia was Manchuria, and chose to pursue a policy of neutrality as the best way of preserving Korean independence as the crisis mounted. In a series of reports to Beijing, Hu Weide, the Chinese ambassador in Saint Petersburg from July 1902 to September 1907, looked closely at whether a Russian or a Japanese victory would be favourable to China, and argued that the latter was preferable, as he maintained a Japanese victory presented the better chance for China to regain sovereignty over Manchuria. In December 1903 China decided to remain neutral if war came, because though Japan was the only power capable of evicting Russia from Manchuria, the extent of Japanese ambitions in Manchuria was not clear to Beijing.
Russian–Japanese negotiations then followed, although by early January 1904 the Japanese government had realised that Russia was not interested in settling the Manchurian or Korean issues. Instead, Russia's goal was buying time—via diplomacy—to further build up militarily. In December 1903, Wilhelm wrote in a marginal note on a diplomatic dispatch about his role in inflaming Russo-Japanese relations:Since 97—Kiaochow—we have never left Russia in any doubt that we would cover her back in Europe, in case she decided to pursue a bigger policy in the Far East that might lead to military complications (with the aim of relieving our eastern border from the fearful pressure and threat of the massive Russian army!). Whereupon, Russia took Port Arthur and trusting us, took her fleet out of the Baltic, thereby making herself vulnerable to us by sea. In Danzig 01 and Reval 02, the same assurance was given again, with result that entire Russian divisions from Poland and European Russia were and are being sent to the Far East. This would not had happened if our governments had not been in agreement!
A recurring theme of Wilhelm's letters to Nicholas was that "Holy Russia" had been "chosen" by God to save the "entire white race" from the "Yellow Peril", and that Russia was "entitled" to annex all of Korea, Manchuria, and northern China up to Beijing. Wilhelm went on to assure Nicholas that once Russia had defeated Japan, this would be a deadly blow to British diplomacy, and that the two emperors, the self-proclaimed "Admiral of the Atlantic" and the "Admiral of the Pacific", would rule Eurasia together, making them able to challenge British sea power as the resources of Eurasia would make their empires immune to a British blockade, and thus allowing Germany and Russia to "divide up the best" of the British colonies in Asia between them.
Nicholas had been prepared to compromise with Japan, but after receiving a letter from Wilhelm attacking him as a coward for his willingness to compromise with the Japanese (who, Wilhelm never ceasing reminding Nicholas, represented the "Yellow Peril") for the sake of peace, became more obstinate. Wilhelm had written to Nicholas stating that the question of Russian interests in Manchuria and Korea was beside the point, saying instead it was a matter of Russiaundertaking the protection and defence of the White Race, and with it, Christian civilization, against the Yellow Race. And whatever the Japs are determined to ensure the domination of the Yellow Race in East Asia, to put themselves at its head and organise and lead it into battle against the White Race. That is the kernel of the situation, and therefore there can be very little doubt about where the sympathies of all half-way intelligent Europeans should lie. England betrayed Europe's interests to America in a cowardly and shameful way over the Panama Canal question, so as to be left in 'peace' by the Yankees. Will the 'Tsar' likewise betray the interests of the White Race to the Yellow as to be 'left in peace' and not embarrass the Hague tribunal too much?.
When Nicholas replied that he still wanted peace, Wilhelm wrote back in a telegram "You innocent angel!", telling his advisors "This is the language of an innocent angel. But not that of a White Tsar!". Nevertheless, Tokyo believed that Russia was not serious about seeking a peaceful solution to the dispute. On 13 January 1904, Japan proposed a formula by which Manchuria would remain outside Japan's sphere of influence and, reciprocally, Korea outside Russia's. On 21 December 1903, the Tarō cabinet voted to go to war against Russia.
By 4 February 1904, no formal reply had been received from Saint Petersburg. On 6 February the Japanese minister to Russia, Kurino Shin'ichirō, was recalled, and Japan severed diplomatic relations with Russia.
Potential diplomatic resolution of territorial concerns between Japan and Russia failed; historians have argued that this directly resulted from the actions of Emperor Nicholas II. Crucially, Nicholas mismanaged his government. Although certain scholars contend that the situation arose from the determination of Nicholas II to use the war against Japan to spark a revival in Russian patriotism, no historical evidence supports this claim. The Tsar's advisors did not support the war, foreseeing problems in transporting troops and supplies from European Russia to the East. The Tsar himself repeatedly delayed negotiations with the Japanese government as he believed that he was protected by God and the autocracy. The Japanese understanding of this can be seen in a telegram from Japanese minister of foreign affairs, Komura, to the minister to Russia, in which he stated:
... the Japanese government have at all times during the progress of the negotiations made it a special point to give prompt answers to all propositions of the Russian government. The negotiations have now been pending for no less than four months, and they have not yet reached a stage where the final issue can with certainty be predicted. In these circumstances the Japanese government cannot but regard with grave concern the situation for which the delays in negotiations are largely responsible.
Some scholars have suggested that Nicholas II dragged Japan into war intentionally, in hopes of reviving Russian nationalism. This notion conflicts with a comment made by Nicholas to Kaiser Wilhelm of Germany, saying there would be no war because he "did not wish it".
This does not reject the claim that Russia played an aggressive role in the East, which it did; rather, it means that Russia unwisely calculated and supposed that Japan would not go to war against Russia's far larger and seemingly superior navy and army. Nicholas held the Japanese in contempt as "yellow monkeys", and he took for granted that the Japanese would simply yield in the face of Russia's superior power, which thus explains his unwillingness to compromise. Evidence of Russia's false sense of security and superiority to Japan is seen by Russian reference to Japan's choosing war as a big mistake.
Declaration of war
Japan issued a declaration of war on 8 February 1904. However, three hours before Japan's declaration of war was received by the Russian government, and without warning, the Japanese Imperial Navy attacked the Russian Far East Fleet at Port Arthur.
Tsar Nicholas II was stunned by news of the attack. He could not believe that Japan would commit an act of war without a formal declaration, and had been assured by his ministers that the Japanese would not fight. When the attack came, according to Cecil Spring Rice, first secretary at the British Embassy, it left the Tsar "almost incredulous".
Russia declared war on Japan eight days later. Japan, in response, made reference to the Russian attack on Sweden in 1808 without declaration of war, although the requirement to mediate disputes between states before commencing hostilities was made international law in 1899, and again in 1907, with the Hague Conventions of 1899 and 1907.
The Qing Empire favoured the Japanese position and even offered military aid, but Japan declined it. However, Yuan Shikai sent envoys to Japanese generals several times to deliver foodstuffs and alcoholic drinks. Native Manchurians joined the war on both sides as hired troops.
Campaign of 1904
Port Arthur, on the Liaodong Peninsula in the south of Manchuria, had been fortified into a major naval base by the Russian Imperial Army. Since it needed to control the sea in order to fight a war on the Asian mainland, Japan's first military objective was to neutralize the Russian fleet at Port Arthur.
Battle of Port Arthur
On the night of 8 February 1904, the Japanese fleet under Admiral Tōgō Heihachirō opened the war with a surprise torpedo boat destroyer attack on the Russian ships at Port Arthur. The attack heavily damaged the and , the heaviest battleships in Russia's Far Eastern theatre, and the 6,600 ton cruiser . These attacks developed into the Battle of Port Arthur the next morning. A series of indecisive naval engagements followed, in which Admiral Tōgō was unable to attack the Russian fleet successfully as it was protected by the shore batteries of the harbour, and the Russians were reluctant to leave the harbour for the open seas, especially after the death of Admiral Stepan Osipovich Makarov on 13 April 1904. Although the actual Battle of Port Arthur was indecisive, the initial attacks had a devastating psychological effect on Russia, which had been confident about the prospect of war. The Japanese had seized the initiative while the Russians waited in port.
These engagements provided cover for a Japanese landing near Incheon in Korea. From Incheon the Japanese occupied Hanseong and then the rest of Korea. After the Japanese occupation of Hanseong, Emperor Gojong sent a detachment of 17,000 soldiers to support Russia. By the end of April, the Japanese Imperial Army under Kuroki Tamemoto was ready to cross the Yalu River into Russian-occupied Manchuria.
Blockade of Port Arthur
The Japanese attempted to deny the Russians use of Port Arthur. During the night of 13–14 February, the Japanese attempted to block the entrance to Port Arthur by sinking several concrete-filled steamers in the deep water channel to the port, but they sank too deep to be effective. A similar attempt to block the harbour entrance during the night of 3–4 May also failed. In March, the charismatic Vice Admiral Makarov had taken command of the First Russian Pacific Squadron with the intention of breaking out of the Port Arthur blockade.
On 12 April 1904, two Russian pre-dreadnought battleships, the flagship and the , slipped out of port but struck Japanese mines off Port Arthur. The Petropavlovsk sank almost immediately, while the Pobeda had to be towed back to port for extensive repairs. Admiral Makarov, the single most effective Russian naval strategist of the war, died on the battleship Petropavlovsk.
On 15 April 1904, the Russian government made overtures threatening to seize the British war correspondents who were taking the ship into war zones to report for the London-based Times newspaper, citing concerns about the possibility of the British giving away Russian positions to the Japanese fleet.
The Russians quickly learned, and soon employed, the Japanese tactic of offensive minelaying. On 15 May 1904, two Japanese battleships, the and the , were lured into a recently laid Russian minefield off Port Arthur, each striking at least two mines. The Hatsuse sank within minutes, taking 450 sailors with her, while the Yashima sank while under tow towards Korea for repairs. On 23 June 1904, a breakout attempt by the Russian squadron, now under the command of Admiral Wilgelm Vitgeft, failed. By the end of the month, Japanese artillery were firing shells into the harbour.
Siege of Port Arthur
The Siege of Port Arthur commenced in April 1904. Japanese troops tried numerous frontal assaults on the fortified hilltops overlooking the harbour, which were defeated with Japanese casualties in the thousands. Eventually, though, with the aid of several batteries of 11-inch (280 mm) howitzers, the Japanese were able to capture the key hilltop bastion in December 1904. With a spotter at the end of a phone line located at this vantage point, the long-range artillery was able to shell the Russian fleet, which was unable to retaliate against the land-based artillery invisible over the other side of hilltop, and was unable or unwilling to sail out against the blockading fleet. Four Russian battleships and two cruisers were sunk in succession, with the fifth and last battleship being forced to scuttle a few weeks later. Thus, all capital ships of the Russian fleet in the Pacific were sunk. This is probably the only example in military history when such a scale of devastation was achieved by land-based artillery against major warships.
Meanwhile, attempts to relieve the besieged city by land also failed, and, after the Battle of Liaoyang in late August, the northern Russian force that might have been able to relieve Port Arthur retreated to Mukden (Shenyang). Major General Anatoly Stessel, commander of the Port Arthur garrison, believed that the purpose of defending the city was lost after the fleet had been destroyed. In general, the Russian defenders were suffering disproportionate casualties each time the Japanese attacked. In particular, several large underground mines were exploded in late December, resulting in the costly capture of a few more pieces of the defensive line. Stessel, therefore, decided to surrender to the surprised Japanese generals on 2 January 1905. He made his decision without consulting either the other military staff present, or the Tsar and military command, all of whom disagreed with the decision. Stessel was convicted by a court-martial in 1908 and sentenced to death on account of an incompetent defense and for disobeying orders. He was later pardoned.
Anglo–Japanese intelligence co-operation
Even before the war, British and Japanese intelligence had co-operated against Russia due to the Anglo-Japanese Alliance. During the war, Indian Army stations in Malaya and China often intercepted and read wireless and telegraph cable traffic relating to the war, which was shared with the Japanese. In their turn, the Japanese shared information about Russia with the British with one British official writing of the "perfect quality" of Japanese intelligence. In particular, British and Japanese intelligence gathered much evidence that Germany was supporting Russia in the war as part of a bid to disturb the balance of power in Europe, which led to British officials increasingly perceiving that country as a threat to the international order.
Battle of Yalu River
In contrast to the Japanese strategy of rapidly gaining ground to control Manchuria, Russian strategy focused on fighting delaying actions to gain time for reinforcements to arrive via the long Trans-Siberian Railway, which was incomplete near Irkutsk at the time. On 1 May 1904, the Battle of Yalu River became the first major land battle of the war; Japanese troops stormed a Russian position after crossing the river. The defeat of the Russian Eastern Detachment removed the perception that the Japanese would be an easy enemy, that the war would be short, and that Russia would be the overwhelming victor. This was also the first battle in decades to be an Asian victory over a European power and marked Russia's inability to match Japan's military prowess. Japanese troops proceeded to land at several points on the Manchurian coast, and in a series of engagements, drove the Russians back towards Port Arthur. The subsequent battles, including the Battle of Nanshan on 25 May 1904, were marked by heavy Japanese losses largely from attacking entrenched Russian positions.
Battle of the Yellow Sea
With the death of Admiral Stepan Makarov during the siege of Port Arthur in April 1904, Admiral Wilgelm Vitgeft was appointed commander of the battle fleet and was ordered to make a sortie from Port Arthur and deploy his force to Vladivostok. Flying his flag in the French-built pre-dreadnought , Vitgeft proceeded to lead his six battleships, four cruisers, and 14 torpedo boat destroyers into the Yellow Sea in the early morning of 10 August 1904. Waiting for him was Admiral Tōgō and his fleet of four battleships, 10 cruisers, and 18 torpedo boat destroyers.
At approximately 12:15, the battleship fleets obtained visual contact with each other, and at 13:00 with Tōgō crossing Vitgeft's T, they commenced main battery fire at a range of about eight miles, the longest ever conducted up to that time. For about thirty minutes the battleships pounded one another until they had closed to less than four miles and began to bring their secondary batteries into play. At 18:30, a hit from one of Tōgō's battleships struck Vitgeft's flagship's bridge, killing him instantly.
With the Tsesarevichs helm jammed and their admiral killed in action, she turned from her battle line, causing confusion among her fleet. However, Tōgō was determined to sink the Russian flagship and continued pounding her, and it was saved only by the gallant charge of the American-built , whose captain successfully drew away Tōgō's heavy fire from the Russian flagship. Knowing of the impending battle with the battleship reinforcements arriving from Russia (the Baltic Fleet), Tōgō chose not to risk his battleships by pursuing his enemy as they turned about and headed back into Port Arthur, thus ending naval history's longest-range gunnery duel up to that time and the first modern clash of steel battleship fleets on the high seas.
Baltic Fleet redeploys
Meanwhile, the Russians were preparing to reinforce their Far East Fleet by sending the Baltic Fleet, under the command of Admiral Zinovy Rozhestvensky. After a false start caused by engine problems and other mishaps, the squadron finally departed on 15 October 1904, and sailed halfway around the world from the Baltic Sea to the Pacific via the Cape Route around the Cape of Good Hope in the course of a seven-month odyssey that was to attract worldwide attention. The Dogger Bank incident on 21 October 1904, where the Russian fleet fired on British fishing boats that they mistook for enemy torpedo boats, nearly sparked a war with the United Kingdom (an ally of Japan, but neutral, unless provoked). During the voyage, the fleet separated into a portion that went through the Suez Canal while the larger battleships went around the Cape of Good Hope.
Effects on civilians
It was reported in 1905 that many Russian females were raped by Japanese troops, causing a widespread infections of venereal diseases in large number of Japanese troops. Hal Gold estimated there twenty thousand recorded incidents of rape against Russian women. During the fighting in Manchuria, there were Russian troops that looted and burned some Chinese villages, raped women and often killed those who resisted or did not understand what they wanted. The Russian justification for all this was that Chinese civilians, being Asian, must have been helping their fellow Asians (the Japanese) inflict defeat on the Russians, and therefore deserved to be punished. The Russian troops were gripped by the fear of the "Yellow Peril", and saw all Asians, not just the Japanese, as the enemy. All of the Russian soldiers were much feared by the Chinese population of Manchuria, but it was the Cossacks whom they feared the most on the account of their brutality and insatiable desire to loot. Largely because of the more disciplined behavior of the Japanese, the Han and Manchu population of Manchuria tended to be pro-Japanese. However Japanese were also prone to looting, albeit in a considerably less brutal manner than the Russians, and summarily executed any Chinese or Manchu whom they suspected of being spies. The city of Liaoyang had the misfortune to be sacked three times within three days: first by the Russians, then by the Chinese police, and finally by the Japanese. The Japanese hired Chinese bandits known variously as the Chunguses, Chunchuse or khunhuzy to engage in guerrilla warfare by attacking Russian supply columns. Only once did the Chunguses attack Japanese forces, and that attack was apparently motivated by the Chunguses mistaking the Japanese forces for a Russian one. Zhang Zuolin, a prominent bandit leader and the future "Old Marshal" who would rule Manchuria as a warlord between 1916 and 1928, worked as a Chunguse for the Japanese. Manchuria was still officially part of the Chinese Empire, and the Chinese civil servants tried their best to be neutral as Russian and Japanese troops marched across Manchuria. In the parts of Manchuria occupied by the Japanese, Tokyo appointed "civil governors" who worked to improve health, sanitation and the state of the roads. These activities were also self-interested, as improved roads lessened Japanese logistics problems while improved health amongst the Chinese lessened the dangers of diseases infecting the Japanese troops. By contrast, the Russians made no effort to improve sanitation or health amongst the Chinese, and destroyed everything when they retreated. Many Chinese tended to see the Japanese as the lesser evil.
Campaign of 1905
With the fall of Port Arthur, the Japanese 3rd Army could continue northward to reinforce positions south of Russian-held Mukden. With the onset of the severe Manchurian winter, there had been no major land engagements since the Battle of Shaho the previous year. The two sides camped opposite each other along 60 to of front lines south of Mukden.
Battle of Sandepu
The Russian Second Army under General Oskar Gripenberg, between 25 and 29 January, attacked the Japanese left flank near the town of Sandepu, almost breaking through. This caught the Japanese by surprise. However, without support from other Russian units the attack stalled, Gripenberg was ordered to halt by Kuropatkin and the battle was inconclusive. The Japanese knew that they needed to destroy the Russian army in Manchuria before Russian reinforcements arrived via the Trans-Siberian railroad.
Battle of Mukden
The Battle of Mukden commenced on 20 February 1905. In the following days Japanese forces proceeded to assault the right and left flanks of Russian forces surrounding Mukden, along a front. Approximately half a million men were involved in the fighting. Both sides were well entrenched and were backed by hundreds of artillery pieces. After days of harsh fighting, added pressure from the flanks forced both ends of the Russian defensive line to curve backwards. Seeing they were about to be encircled, the Russians began a general retreat, fighting a series of fierce rearguard actions, which soon deteriorated in the confusion and collapse of Russian forces. On 10 March 1905, after three weeks of fighting, General Kuropatkin decided to withdraw to the north of Mukden. The Russians suffered 90,000 casualties in the battle.
The retreating Russian Manchurian Army formations disbanded as fighting units, but the Japanese failed to destroy them completely. The Japanese themselves had suffered heavy casualties and were in no condition to pursue. Although the Battle of Mukden was a major defeat for the Russians and was the most decisive land battle ever fought by the Japanese, the final victory still depended on the navy.
Battle of Tsushima
After a stopover of several weeks at the minor port of Nossi-Bé, Madagascar, that had been reluctantly allowed by neutral France in order not to jeopardize its relations with its Russian ally, the Russian Baltic fleet proceeded to Cam Ranh Bay in French Indochina passing on its way through the Singapore Strait between 7 and 10 April 1905. The fleet finally reached the Sea of Japan in May 1905. The logistics of such an undertaking in the age of coal power was astounding. The squadron required approximately 500,000 tons of coal to complete the journey, yet by international law, it was not allowed to coal at neutral ports, forcing the Russian authorities to acquire a large fleet of colliers to supply the fleet at sea. The weight of the ships' stores needed for such a long journey was to be another major problem.
The Russian Second Pacific Squadron (the renamed Baltic Fleet) sailed 18,000 nautical miles (33,000 km) to relieve Port Arthur only to hear the demoralizing news that Port Arthur had fallen while it was still at Madagascar. Admiral Rozhestvensky's only hope now was to reach the port of Vladivostok. There were three routes to Vladivostok, with the shortest and most direct passing through Tsushima Strait between Korea and Japan. However, this was also the most dangerous route as it passed between the Japanese home islands and the Japanese naval bases in Korea.
Admiral Tōgō was aware of Russian progress and understood that, with the fall of Port Arthur, the Second and Third Pacific squadrons would try to reach the only other Russian port in the Far East, Vladivostok. Battle plans were laid down and ships were repaired and refitted to intercept the Russian fleet.
The Japanese Combined Fleet, which had originally consisted of six battleships, was now down to four battleships and one second class battleship (two had been lost to mines), but still retained its cruisers, destroyers, and torpedo boats. The Russian Second Pacific Squadron contained eight battleships, including four new battleships of the , as well as cruisers, destroyers and other auxiliaries for a total of 38 ships.
By the end of May, the Second Pacific Squadron was on the last leg of its journey to Vladivostok, taking the shorter, riskier route between Korea and Japan, and travelling at night to avoid discovery. Unfortunately for the Russians, while in compliance with the rules of war, the two trailing hospital ships had continued to burn their lights, which were spotted by the Japanese armed merchant cruiser Shinano Maru. Wireless communication was used to inform Togo's headquarters, where the Combined Fleet was immediately ordered to sortie. Still receiving reports from scouting forces, the Japanese were able to position their fleet to "cross the T" of the Russian fleet. The Japanese engaged the Russians in the Tsushima Straits on 27–28 May 1905. The Russian fleet was virtually annihilated, losing eight battleships, numerous smaller vessels, and more than 5,000 men, while the Japanese lost three torpedo boats and 116 men. Only three Russian vessels escaped to Vladivostok, while six others were interned in neutral ports. After the Battle of Tsushima, a combined Japanese Army and Navy operation occupied Sakhalin Island to force the Russians into suing for peace.
Peace and aftermath
Treaty of Portsmouth
Military leaders and senior tsarist officials agreed before the war that Russia was a much stronger nation and had little to fear from the Oriental newcomer. The fanatical zeal of the Japanese infantrymen astonished the Russians, who were dismayed by the apathy, backwardness, and defeatism of their own soldiers. The defeats of the Army and Navy shook up Russian confidence. Throughout 1905, the Imperial Russian government was rocked by revolution. The population was against escalation of the war. The empire was certainly capable of sending more troops but this would make little difference in the outcome due to the poor state of the economy, the embarrassing defeats of the Russian Army and Navy by the Japanese, and the relative unimportance to Russia of the disputed land made the war extremely unpopular. Tsar Nicholas II elected to negotiate peace so he could concentrate on internal matters after the disaster of Bloody Sunday on 9 January 1905.
Both sides accepted the offer of United States President Theodore Roosevelt to mediate. Meetings were held in Portsmouth, New Hampshire, with Sergei Witte leading the Russian delegation and Baron Komura leading the Japanese delegation. The Treaty of Portsmouth was signed on 5 September 1905 at the Portsmouth Naval Shipyard. Witte became Russian Prime Minister the same year.
After courting the Japanese, Roosevelt decided to support the Tsar's refusal to pay indemnities, a move that policymakers in Tokyo interpreted as signifying that the United States had more than a passing interest in Asian affairs. Russia recognized Korea as part of the Japanese sphere of influence and agreed to evacuate Manchuria. Japan would annex Korea in 1910 (Japan–Korea Treaty of 1910), with scant protest from other powers. From 1910 forward, the Japanese adopted a strategy of using the Korean Peninsula as a gateway to the Asian continent and making Korea's economy subordinate to Japanese economic interests.
Russia also signed over its 25-year leasehold rights to Port Arthur, including the naval base and the peninsula around it, and ceded the southern half of Sakhalin Island to Japan. Sakhalin would be taken back by the Soviet Union following the defeat of the Japanese in World War II.
Roosevelt earned the Nobel Peace Prize for his effort. George E. Mowry concludes that Roosevelt handled the arbitration well, doing an "excellent job of balancing Russian and Japanese power in the Orient, where the supremacy of either constituted a threat to growing America". As Japan had won every battle on land and sea and as the Japanese people did not understand that the costs of the war had pushed their nation to the verge of bankruptcy, the Japanese public was enraged by the Treaty of Portsmouth as many Japanese had expected the war to end with Russia ceding the Russian Far East to Japan and for Russia to pay an indemnity. The United States was widely blamed in Japan for the Treaty of Portsmouth with Roosevelt having allegedly "cheated" Japan out of its rightful claims at the peace conference. On 5 September 1905 the Hibiya incendiary incident - as the anti-American riots were euphemistically described - erupted in Tokyo and lasted for three days, forcing the government to declare martial law.
Casualties
Sources do not agree on a precise number of deaths from the war because of a lack of body counts for confirmation. The number of Japanese Army dead in combat or died of wounds is put at around 59,000 with around 27,000 additional casualties from disease, and between 6,000 and 12,000 wounded. Estimates of Russian Army dead range from around 34,000 to around 53,000 men with a further 9,000–19,000 dying of disease and around 75,000 captured. The total number of dead for both sides is generally stated as around 130,000 to 170,000. China suffered 20,000 civilian deaths, and financially the loss amounted to over 69 million taels' worth of silver.
During many of the battles at sea, several thousand soldiers being transported drowned after their ships went down. There was no consensus about what to do with transported soldiers at sea, and as a result, many of the ships failed or refused to rescue soldiers that were left shipwrecked. This led to the creation of the second Geneva Convention in 1906, which gave protection and care for shipwrecked soldiers in armed conflict.
Political consequences
This was the first major military victory in the modern era of an Asian power over a European nation. Russia's defeat was met with shock in the West and across the Far East. Japan's prestige rose greatly as it came to be seen as a modern nation. Concurrently, Russia lost virtually its entire Pacific and Baltic fleets, and also much international esteem. This was particularly true in the eyes of Germany and Austria-Hungary before World War I. Russia was France's and Serbia's ally, and that loss of prestige had a significant effect on Germany's future when planning for war with France, and in supporting Austria-Hungary's war with Serbia.
In the absence of Russian competition, and with the distraction of European nations during World War I, combined with the Great Depression that followed, the Japanese military began efforts to dominate China and the rest of Asia, which eventually led to the Second Sino-Japanese War and the Pacific War theatres of World War II.
Effects on Russia
Though there had been popular support for the war among the Russian public following the Japanese attack at Port Arthur in 1904, that popular support soon turned to discontent after suffering multiple defeats at the hands of the Japanese forces. For many Russians, the immediate shock of unexpected humiliation at the hands of Japan caused the conflict to be viewed as a metaphor for the shortcomings of the Romanov autocracy. Popular discontent in Russia after the war added more fuel to the already simmering Russian Revolution of 1905, an event Nicholas II had hoped to avoid entirely by taking intransigent negotiating stances prior to coming to the table. Twelve years later, that discontent boiled over into the February Revolution of 1917. In Poland, which Russia partitioned in the late 18th century, and where Russian rule already caused two major uprisings, the population was so restless that an army of 250,000–300,000—larger than the one facing the Japanese—had to be stationed to put down the unrest. Some political leaders of the Polish insurrection movement (in particular, Józef Piłsudski) sent emissaries to Japan to collaborate on sabotage and intelligence gathering within the Russian Empire and even plan a Japanese-aided uprising.
In Russia, the defeat of 1905 led in the short term to a reform of the Russian military that allowed it to face Germany in World War I. However, the revolts at home following the war planted seeds that presaged the Russian Revolution of 1917. This was because Tsar Nicholas II issued the October Manifesto, which included only limited reforms such as the Duma and failed to address the societal problems of Russia at the time.
Effects on Japan
Japan had become the rising Asian power and had proven that its military could combat the major powers in Europe with success. Most Western powers were stunned that the Japanese not only prevailed but decisively defeated Russia. In the Russo-Japanese War, Japan had also portrayed a sense of readiness in taking a more active and leading role in Asian affairs, which in turn had led to widespread nationalism throughout the region.
Although the war had ended in a victory for Japan, Japanese public opinion was shocked by the very restrained peace terms which were negotiated at the war's end. Widespread discontent spread through the populace upon the announcement of the treaty terms. Riots erupted in major cities in Japan. Two specific requirements, expected after such a costly victory, were especially lacking: territorial gains and monetary reparations to Japan. The peace accord led to feelings of distrust, as the Japanese had intended to retain all of Sakhalin Island, but were forced to settle for half of it after being pressured by the United States, with President Roosevelt opting to support Nicholas II's stance on not ceding territory or paying reparations. The Japanese had wanted reparations to help families recover from lost fathers and sons as well as heavy taxation from the government. Without them, they were at a loss.
The U.S held strength in the Asian region from aggravating European imperialist encroachment. To Japan, this represented a developing threat to the autonomy of the region. U.S.–Japanese relations would recover a bit in the early 20th century, but by the early 1920s, few in Japan believed that the United States meant anything positive for the future of Asia. By the 1930s, the U.S. presence in Asian affairs, along with the instability in China and the collapse of the Western economic order, Japan would act aggressively with respect to China, setting the precedent that would ultimately culminate in the Greater East Asia Co-Prosperity Sphere. Some scholars suggest that Japan's road to World War II had begun not upon winning the Russo-Japanese War, but when it lost the peace.
Assessment
Historical significance
The effects and impact of the Russo-Japanese War introduced a number of characteristics that came to define 20th-century politics and warfare. Many of the innovations brought by the Industrial Revolution, such as rapid-firing artillery and machine guns, as well as more accurate rifles, were first tested on a mass scale then. Military operations on both sea and land showed that modern warfare had undergone a considerable change since the Franco-Prussian War of 1870–71. Most army commanders had previously envisioned using these weapon systems to dominate the battlefield on an operational and tactical level but, as events played out, the technological advances forever altered the conditions of war too.
For East Asia this was the first confrontation after thirty years involving two modern armed forces.
The advanced weaponry led to massive casualty counts. Neither Japan nor Russia had prepared for the number of deaths that would occur in this new kind of warfare, or had the resources to compensate for such losses. This also left its impression on society at large, with the emergence of transnational and nongovernmental organizations, like the Red Cross, becoming prominent after the war. The consequent identification of common problems and challenges began the slow process that came to dominate much of the 20th century.
It has also been argued that the conflict had characteristics of what was later to be described as "total war". These included the mass mobilization of troops into battle and the need for so extensive a supply of equipment, armaments, and supplies that both domestic support and foreign aid were required. It is also argued that domestic response in Russia to the inefficiencies of the tsarist government set in motion the eventual dissolution of the Romanov dynasty.
Reception around the world
To the Western powers, Japan's victory demonstrated the emergence of a new Asian regional power. With the Russian defeat, some scholars have argued that the war had set in motion a change in the global world order with the emergence of Japan as not only a regional power, but rather, the main Asian power. Rather more than the possibilities of diplomatic partnership were emerging, however. The US and Australian reaction to the changed balance of power brought by the war was mixed with fears of a Yellow Peril eventually shifting from China to Japan. American figures such as W. E. B. Du Bois and Lothrop Stoddard saw the victory as a challenge to western supremacy. This was reflected in Austria, where Baron Christian von Ehrenfels interpreted the challenge in racial as well as cultural terms, arguing that "the absolute necessity of a radical sexual reform for the continued existence of the western races of men has ... been raised from the level of discussion to the level of a scientifically proven fact". To stop the Japanese "Yellow Peril" would require drastic changes to society and sexuality in the West.
Certainly the Japanese success increased self-confidence among anti-colonial nationalists in colonised Asian countries – Vietnamese, Indonesians, Indians and Filipinos – and to those in declining countries like the Ottoman Empire and Persia in immediate danger of being absorbed by the Western powers. It also encouraged the Chinese who, despite having been at war with the Japanese only a decade before, still considered Westerners the greater threat. As Sun Yat-sen commented, "We regarded that Russian defeat by Japan as the defeat of the West by the East. We regarded the Japanese victory as our own victory". Even in far-off Tibet the war was a subject of conversation when Sven Hedin visited the Panchen Lama in February 1907. While for Jawaharlal Nehru, then only an aspiring politician in British India, "Japan's victory lessened the feeling of inferiority from which most of us suffered. A great European power had been defeated, thus Asia could still defeat Europe as it had done in the past." And in the Ottoman Empire too, the Committee of Union and Progress embraced Japan as a role model.
In Europe, subject populations were similarly encouraged. James Joyce's novel Ulysses, set in Dublin in 1904, contains hopeful Irish allusions as to the outcome of the war. And in partitioned Poland the artist Józef Mehoffer chose 1905 to paint his "Europa Jubilans" (Europe rejoicing), which portrays an aproned maid taking her ease on a sofa against a background of Eastern artefacts. Painted following demonstrations against the war and Russian cultural suppression, and in the year of Russia's defeat, its subtly coded message looks forward to a time when the Tsarist masters will be defeated in Europe as they had been in Asia.
The significance of the war for oppressed classes as well as subject populations was clear too to Socialist thinkers.
It was this realisation of the universal significance of the war that underlines the historical importance of the conflict and its outcome.
Military results
Russia had lost two of its three fleets. Only its Black Sea Fleet remained, and this was the result of an earlier treaty that had prevented the fleet from leaving the Black Sea. Japan became the sixth-most powerful naval force by combined tonnage, while the Russian Navy declined to one barely stronger than that of Austria–Hungary. The actual costs of the war were large enough to affect the Russian economy and, despite grain exports, the nation developed an external balance of payments deficit. The cost of military re-equipment and re-expansion after 1905 pushed the economy further into deficit, although the size of the deficit was obscured.
The Japanese were on the offensive for most of the war and used massed infantry assaults against defensive positions, which would later become the standard of all European armies during World War I. The battles of the Russo-Japanese War, in which machine guns and artillery took a heavy toll on Russian and Japanese troops, were a precursor to the trench warfare of World War I. A German military advisor sent to Japan, Jakob Meckel, had a tremendous impact on the development of the Japanese military training, tactics, strategy, and organization. His reforms were credited with Japan's overwhelming victory over China in the First Sino-Japanese War of 1894–1895. However, his over-reliance on infantry in offensive campaigns also led to a large number of Japanese casualties.
Military and economic exhaustion affected both countries. Japanese historians regard this war as a turning point for Japan, and a key to understanding the reasons why Japan may have failed militarily and politically later. After the war, acrimony was felt at every level of Japanese society and it became the consensus within Japan that their nation had been treated as the defeated power during the peace conference. As time went on, this feeling, coupled with the sense of "arrogance" at becoming a Great Power, grew and added to growing Japanese hostility towards the West, and fuelled Japan's military and imperial ambitions. Furthermore, Japan's substantiated interests in Korea and Liaodong led to the creation of a Kwantung Army, which became an autonomous and increasingly powerful regional force. Only five years after the war, Japan de jure annexed Korea as part of its colonial empire. Two decades after that, the Kwantung Army staged an incident that led to the invasion of Manchuria in the Mukden Incident; the Kwantung Army eventually came to be heavily involved in the state's politics and administration, leading to a series of localized conflicts with Chinese regional warlords that finally extended into the Second Sino-Japanese War in 1937. As a result, most Chinese historians consider the Russo-Japanese War as a key development in Japan's spiral into militarism in the 1920s–30s.
Following the victory of the Battle of Tsushima, Japan's erstwhile British ally presented a lock of Admiral Nelson's hair to the Imperial Japanese Navy, judging its performance then as on a par with Britain's victory at Trafalgar in 1805. It is still on display at Kyouiku Sankoukan, a public museum maintained by the Japan Self-Defence Force. Nevertheless, there was a consequent shift in British strategic thinking, resulting in enlargement of its naval docks at Auckland, New Zealand; Bombay, British India; Fremantle and Sydney, Australia; Simon's Town, Cape Colony; Singapore and British Hong Kong. The naval war confirmed the direction of the British Admiralty's thinking in tactical terms even as it undermined its strategic grasp of a changing world. Tactical orthodoxy, for example, assumed that a naval battle would imitate the conditions of stationary combat and that ships would engage in one long line sailing on parallel courses; but more flexible tactical thinking would now be required as a firing ship and its target manoeuvred independently.
Military attachés and observers
Military and civilian observers from every major power closely followed the course of the war. Most were able to report on events from the perspective of embedded positions within the land and naval forces of both Russia and Japan. These military attachés and other observers prepared first-hand accounts of the war and analytical papers. In-depth observer narratives of the war and more narrowly focused professional journal articles were written soon after the war; and these post-war reports conclusively illustrated the battlefield destructiveness of this conflict. This was the first time the tactics of entrenched positions for infantry defended with machine guns and artillery became vitally important. Both would become dominant factors in World War I. Even though entrenched positions had already been a significant part of both the Franco-Prussian War and the American Civil War, it is now apparent that the high casualty counts, and the tactical lessons readily available to observer nations, were completely disregarded in preparations for war in Europe, and during much of the course of World War I.
In 1904–1905, Ian Standish Monteith Hamilton was the military attaché of the British Indian Army serving with the Imperial Japanese Army in Manchuria. As one of the several military attachés from Western countries, he was the first to arrive in Japan after the start of the war.
He therefore would be recognized as the dean of multi-national attachés and observers in this conflict, although out-ranked by British field marshal, William Gustavus Nicholson, 1st Baron Nicholson, who was later to become chief of the Imperial General Staff.
Financing
Despite its gold reserves of 106.3 million pounds, Russia's pre-war financial situation was not enviable. The country had large budget deficits year after year, and was largely dependent on borrowed money.
Russia's war effort was funded primarily by France, in a series of loans totalling 800 million francs (30.4 million pounds); another loan in the amount of 600 million francs was agreed upon, but later cancelled. These loans were extended within a climate of mass bribing of the French press (made necessary by Russia's precarious economic and social situation and poor military performance). Although initially reluctant to participate in the war, the French government and major banks were co-operative since it became clear that Russian and French economic interests were tied. In addition to French money, Russia secured a loan in the amount of 500 million marks (24.5 million pounds) from Germany, who also financed Japan's war effort.
Conversely, Japan's pre-war gold reserves were a modest 11.7 million pounds; a major portion of the total cost of the war was covered by money borrowed from the United Kingdom, Canada, and the United States.
During his canvassing expedition in London, the Japanese vice-governor of the Bank of Japan Takahashi Korekiyo met Jacob Schiff, an American banker and head of Kuhn, Loeb & Co. Schiff, in response to Russia's anti-Jewish pogroms and sympathetic to Japan's cause, extended a critical series of loans to the Empire of Japan, in the amount of 200 million US dollars (41.2 million pounds). He also raised loans from the Rothschild family in Britain.
Japan's total war expenditure was 2,150 million yen, of which 38%, or 820 million yen, was raised overseas.
List of battles
8 February 1904: Battle of Port Arthur: naval battle, inconclusive
9 February 1904: Battle of Chemulpo Bay: naval battle, Japanese victory
30 April-1 May 1904: Battle of Yalu River, Japanese victory
25–26 May 1904: Battle of Nanshan: Japanese victory
14–15 June 1904: Battle of Te-li-Ssu: Japanese victory
17 July 1904: Battle of Motien Pass: Japanese victory
24 July 1904: Battle of Tashihchiao: Japanese victory
31 July 1904: Battle of Hsimucheng: Japanese victory
10 August 1904: Battle of the Yellow Sea,: naval battle, Japanese victory
14 August 1904: Battle off Ulsan: naval battle, Japanese victory
20 August 1904: Battle of Korsakov: naval battle, Japanese victory
19 August 1904 – 2 January 1905: Siege of Port Arthur, Japanese victory
25 August-3 September 1904: Battle of Liaoyang: Japanese victory
5–17 October 1904 Battle of Shaho: Inconclusive
26–27 January 1905: Battle of Sandepu: Inconclusive
21 February-10 March 1905: Battle of Mukden: Japanese victory
27–28 May 1905: Battle of Tsushima: naval battle, Japanese victory
7–31 July 1905: Invasion of Sakhalin: Japanese victory
Cultural legacy
Visual arts
The Russo-Japanese War was covered by dozens of foreign journalists who sent back sketches that were turned into lithographs and other reproducible forms. Propaganda images were circulated by both sides, often in the form of postcards and based on insulting racial stereotypes. These were produced not only by the combatants but by those from European countries who supported one or the other side or had a commercial or colonial stake in the area. War photographs were also popular, appearing in both the press and in book form.
In Russia, the war was covered by anonymous satirical graphic luboks for sale in markets, recording the war for the domestic audience. Around 300 were made before their creation was banned by the Russian government. Their Japanese equivalents were woodblock prints. These had been common during the Sino-Japanese war a decade earlier and celebrations of the new conflict tended to repeat the same imagery and situations. But by this time in Japan postcards had become the most common form of communication and they soon replaced prints as a medium for topographical imagery and war reportage. In some ways, however, they were still dependent on the print for their pictorial conventions, not least in issuing the cards in series that assembled into a composite scene or design, either as diptychs, triptychs or even more ambitious formats. However, captioning swiftly moved from the calligraphic side inscription to a printed title below, and not just in Japanese but in English and other European languages. There was a lively sense that these images served not only as mementoes but also as propaganda statements.
War artists were to be found on the Russian side and even figured among the casualties. Vasily Vereshchagin went down with the Petropavlovsk, Admiral Makarov's flagship, when it was sunk by mines. However, his last work, a picture of a council of war presided over by the admiral, was recovered almost undamaged. Another artist, Mykola Samokysh, first came to notice for his reports during the war and the paintings worked up from his diary sketch-books. Other depictions appeared after the event. The two by the Georgian naïve painter Niko Pirosmani from 1906 must have been dependent on newspaper reports since he was not present. Then, in 1914 at the outset of World War I, Yury Repin made an episode during the Battle of Yalu River the subject of a broad heroic canvas.
Music
On either side, there were lyrics lamenting the necessity of fighting in a foreign land, far from home. One of the earliest of several Russian songs still performed today was the waltz "Amur's Waves" (Amurskie volny), which evokes the melancholy of standing watch on the motherland far east frontier.
Two others grew out of incidents during the war. "On the Hills of Manchuria" (Na sopkah Manchzhurii; 1906) is another waltz composed by Ilya Shatrov, a decorated military musician whose regiment suffered badly in the Battle of Mukden. Originally only the music was published, and the words by Stepan Petrov were added later.
The second song, "Variag", commemorates the Battle of Chemulpo Bay in which that cruiser and the gunboat Korietz steamed out to confront an encircling Japanese squadron rather than surrender. That act of heroism was first celebrated in a German song by Rudolf Greintz in 1907, which was quickly translated into Russian and sung to a martial accompaniment. These lyrics mourned the fallen lying in their graves and threatened revenge.
Nikolai Rimsky-Korsakov also reacted to the war by composing the satirical opera The Golden Cockerel, completed in 1907. Although it was ostensibly based on a verse fairy tale by Alexander Pushkin written in 1834, the authorities quickly realised its true target and immediately banned it from performance. The opera was premiered in 1909, after Rimsky-Korsakov's death, and even then with modifications required by the censors.
Poetry
Some Japanese poetry dealing with the war still has a high profile. General Nogi Maresuke's "Outside the Goldland fortress" was learned by generations of schoolchildren and valued for its bleak stoicism. The army surgeon Mori Ōgai kept a verse diary which tackled such themes as racism, strategic mistakes and the ambiguities of victory which can now be appreciated in historical hindsight. Nowadays too there is growing appreciation of Yosano Akiko's parting poem to her brother as he left for the war, which includes the critical lines.
Even the Emperor Meiji himself entered the poetic lists, writing in answer to all the lamentations about death in a foreign land that the patriotic soul returns to the homeland.
European treatments were similarly varied. Jane H. Oakley attempted an epic treatment of the conflict in 86 cantos. The French poet Blaise Cendrars was later to represent himself as on a Russian train on its way to Manchuria at the time in his La prose du Transsibérien et de la Petite Jehanne de France (1913) and energetically evoked the results of the war along the way:
Much later, the Scottish poet Douglas Dunn devoted an epistolary poem in verse to the naval war in The Donkey's Ears: Politovsky's Letters Home (2000). This follows the voyage of the Russian Imperial Navy flagship Kniaz to its sinking at the Battle of Tsushima.
Fiction
Fictional coverage of the war in English began even before it was over. An early example was Allen Upward's The International Spy. Set in both Russia and Japan, it ends with the Dogger Bank incident involving the Baltic Fleet. The political thinking displayed there is typical of the time. There is great admiration for the Japanese, who were British allies. Russia is in turmoil, but the main impetus towards war is not imperialism as such but commercial forces. "Every student of modern history has remarked the fact that all recent wars have been promoted by great combinations of capitalists. The causes which formerly led to war between nation and nation have ceased to operate" (p. 40). The true villain plotting in the background, however, is the German Emperor, seeking to destabilise the European balance of power in his country's favour. Towards the end of the novel, the narrator steals a German submarine and successfully foils a plot to involve the British in the war. The submarine motif reappeared in George Griffith's science fiction novel, The Stolen Submarine (1904), although in this case it is a French super-submarine which its developer sells to the Russians for use against the Japanese in another tale of international intrigue.
Though most English-language fiction of the period took the Japanese side, the Rev. W. W. Walker's Canadian novella, Alter Ego, is an exception. It features a Canadian volunteer in the Russian army who, on his return, agrees to talk about his experiences to an isolated upcountry community and relates his part in the Battle of Mukden. Though this incident only occupies two of the book's six chapters, it is used to illustrate the main message there, that war is "anti-Christian and barbarous, except in a defensive sense" (Ch.3).
Various aspects of the war were also common in contemporary children's fiction. Categorised as Boys' Own adventure stories, they offer few insights into the conflict, being generally based on news articles and sharing without any reflection in the contemporary culture of imperialism. Among these, Herbert Strang was responsible for two novels: Kobo told from the Japanese side, and Brown of Moukden viewed from the Russian side. Three more were written by the prolific American author, Edward Stratemeyer: Under the Mikado's Flag, At the Fall of Port Arthur, and Under Togo for Japan, or Three Young Americans on Land and Sea (1906). Two other English-language stories begin with the action at Port Arthur and follow the events thereafter: A Soldier of Japan: a tale of the Russo-Japanese War by Captain Frederick Sadleir Brereton, and The North Pacific by Willis Boyd Allen (1855–1938). Two more also involve young men fighting in the Japanese navy: Americans in For the Mikado by Kirk Munroe, and a temporarily disgraced English officer in Under the Ensign of the Rising Sun by Harry Collingwood, the pen-name of William Joseph Cosens Lancaster (1851–1922), whose speciality was naval fiction.
Another literary genre affected by the outcome of the war was invasion literature, either fuelled by racialist fears or generated by the international power struggle. Shunrō Oshikawa's novel The Submarine Battleship (Kaitei Gunkan) was published in 1900 before the actual fighting began but shared the imperial tensions that produced it. It is the story of an armoured ram-armed submarine involved in a Russo-Japanese conflict. Three other novels appeared in 1908 and are thought of as significant now because of their prophetic dimension. American author Arthur Wellesley Kipling (1885–1947) prefaced his The New Dominion – A Tale of Tomorrow's Wars with a note counselling future vigilance. The scenario there is an attack by German and Japanese allies which the US and British navies victoriously fend off. In Germany itself an air attack on the American fleet is described by Ferdinand Heinrich Grautoff (1871–1935), writing under the name Parabellum, in his novel Banzai!. Published in Berlin in 1908, it was translated into English the following year. An Australian author using the pseudonym Charles H. Kirmess first serialised his The Commonwealth Crisis and then revised it for book publication as The Australian Crisis in 1909. It is set in 1912 and told from the standpoint of 1922, following a military invasion of Australia's Northern Territory and colonisation by Japanese settlers.
Most Russian fictional accounts of the war had a documentary element. Alexey Novikov-Priboy served in the Baltic Fleet and wrote about the conflict on his return, but his early work was suppressed. It was not until the changed political climate under Soviet rule that he began writing his historical epic Tsushima, based on his own experiences on board the battleship as well as on testimonies of fellow sailors and government archives. The first part was published in 1932, the second in 1935, and the whole novel was later awarded the Stalin Prize. It describes the heroism of Russian sailors and certain officers whose defeat, in accordance with the new Soviet thinking, was due to the criminal negligence of the Imperial Naval command. A German novel by Frank Thiess, originally published as Tsushima in 1936 (and later translated as The Voyage of Forgotten Men), covered the same journey round the world to defeat.
Later there appeared a first-hand account of the siege of Port Arthur by Alexander Stepanov (1892–1965). He had been present there as the 12-year-old son of a battery commander and his novel, Port Arthur: a historical narrative (1944), is based on his own diaries and his father's notes. The work is considered one of the best historical novels of the Soviet period. A later novel in which the war appears is Valentin Pikul's The Three Ages of Okini-San (1981). Centred on the life of Vladimir Kokovtsov, who rose through the ranks to admiral of the Russian fleet, it covers the period from the Russo-Japanese War through to the February and October Revolutions. A much later Russian genre novel uses the period of the war as background. This is Boris Akunin's The Diamond Chariot (2003), in the first part of which the detective Erast Fandorin is charged with protecting the Trans-Siberian Railway from Japanese sabotage.
The main historical novel dealing with the war from the Japanese side is Shiba Ryōtarō's Clouds Above the Hill, published serially in several volumes between 1968 and 1972, and translated in English in 2013. The closely researched story spans the decade from the Sino-Japanese War to the Russo-Japanese War and went on to become the nation's favourite book.
Filmography
Port Arthur (1936)
Kreiser Varyag (1946)
Nichiro sensō shōri no hishi: Tekichū ōdan sanbyaku-ri (1957)
Meiji tennô to nichiro daisenso (1958)
The Battle of the Japan Sea (1969, , Nihonkai-Daikaisen) depicts the naval battles of the war, the attacks on the Port Arthur highlands, and the subterfuge and diplomacy of Japanese agents in Sweden. Admiral Togo is portrayed by Toshiro Mifune.
The Battle of Tsushima (1975) documentary, depiction of the naval Battle of Tsushima
The Battle of Port Arthur (1980, sometimes referred as 203 Kochi), depiction of the Siege of Port Arthur
Nihonkai daikaisen: Umi yukaba (1983)
Reilly, Ace of Spies (1983). Russian-born British spy Sidney Reilly's role in providing intelligence that allowed the Japanese surprise attack that started the Siege of Port Arthur is dramatised in the second episode of this TV series.
Saka no Ue no Kumo, of which the third series dealt with the war period (December 2011)
Golden Kamuy (2018), an anime adaptation of the manga of the same name. The story takes place just after the Russo-Japanese War, and features many flashbacks and references to it. Several of its principal characters are Japanese army men who fought in the Siege of Port Arthur.
The Prisoner of Sakura (2019, , Sorokin no mita saka, , V plenu u sakury), a joint Japan-Russia co-production, which was based on the true story of a prison camp in Matsuyama, Ehime Prefecture, during the war. The movie is centered around the romance between a Russian officer, a prisoner of war, and a Japanese nurse who find themselves on opposing sides of the war.
See also
Foreign policy of the Russian Empire
Kaneko Kentarō
Outline of war
List of warships sunk during the Russo-Japanese War
Manchuria under Qing rule
Baron Rosen
Russian Imperialism in Asia and the Russo-Japanese War
Western imperialism in Asia
References
Bibliography
online
In
Kowner, Rotem. Historical Dictionary of the Russo-Japanese War, also published as The A to Z of the Russo-Japanese War (2009) excerpt
In .
In .
Further reading
Corbett, Sir Julian. Maritime Operations in the Russo-Japanese War 1904–1905. (1994) Originally classified, and in two volumes, .
Dower, John W., Throwing off Asia III, Woodblock prints of the Russo-Japanese War, 2008, MIT Visualizing Cultures
Hall, Richard C. "The Next War: The Influence of the Russo-Japanese War on Southeastern Europe and the Balkan Wars of 1912–1913." Journal of Slavic Military Studies 17.3 (2004): 563–577.
Hough, Richard A. The Fleet That Had To Die. Ballantine Books. (1960).
Japanese Ministry of Foreign Affairs, Correspondence Regarding the Negotiations between Japan and Russia (1903–1904), Presented to the Imperial Diet, March 1904 (Tokyo, 1904)
Jentschura, Hansgeorg; Dieter Jung, Peter Mickel. Warships of the Imperial Japanese Navy, 1869–1945. United States Naval Institute, Annapolis, Maryland, 1977. Originally published in German as Die Japanischen Kreigschiffe 1869–1945 in 1970, translated into English by David Brown and Antony Preston. .
Kowner, Rotem (2006). Historical Dictionary of the Russo-Japanese War. Scarecrow. .
Matsumura Masayoshi, Ian Ruxton (trans.), Baron Kaneko and the Russo-Japanese War (1904–05), Lulu Press 2009
Murray, Nicholas (2013). The Rocky Road to the Great War: the Evolution of Trench Warfare to 1914. Dulles, Virginia, Potomac Books
Morris, Edmund (2002). Theodore Rex, Google Books. New York: Random House.
Nish, Ian Hill. (1985). The Origins of the Russo-Japanese War. London: Longman.
Novikov-Priboy, Aleksei. Tsushima. (An account from a seaman aboard the , which was captured at Tsushima). London: George Allen & Unwin Ltd. (1936).
Okamoto, Shumpei (1970). The Japanese Oligarchy and the Russo-Japanese War. Columbia University Press.
Pleshakov, Constantine. The Tsar's Last Armada: The Epic Voyage to the Battle of Tsushima. . (2002).
Podalko, Petr E. "'Weak ally'or 'strong enemy?': Japan in the eyes of Russian diplomats and military agents, 1900–1907." Japan Forum 28#3 (2016).
Tomitch, V. M. Warships of the Imperial Russian Navy. Volume 1, Battleships. (1968).
Illustrations
Gerbig-Fabel, Marco. "Photographic artefacts of war 1904–1905: the Russo-Japanese war as transnational media event." European Review of History—Revue européenne d'histoire 15.6 (2008): 629–642.
Saaler, Sven und Inaba Chiharu (Hg.). Der Russisch-Japanische Krieg 1904/05 im Spiegel deutscher Bilderbogen, Deutsches Institut für Japanstudien Tokyo, (2005).
Sharf, Frederick A. and James T. Ulak, eds. A Well-Watched War: Images from the Russo-Japanese Front, 1904–1905 (Newbury, MA, 2000), the catalogue of the show at the Sackler Gallery in Washington, DC,
Historiography
Hamby, Joel E. "Striking the Balance: Strategy and Force in the Russo-Japanese War." Armed Forces & Society 30.3 (2004): 325–356.
Seager, Robert. Alfred Thayer Mahan: The Man And His Letters. (1977) .
van der Oye, David Schimmelpenninck. "Rewriting the Russo-Japanese War: A Centenary Retrospective." The Russian Review 67.1 (2008): 78–87. online
Won-soo, Kim. "Trends in the Study of the Russo-Japanese War in Korea and Future Tasks-Third-party perspective on the origins of the war." International Journal of Korean History 7 (2005): 1–28. online
External links
RussoJapaneseWar.com , Russo-Japanese War research society.
BFcollection.net, Database of Russian Army Jewish soldiers injured, killed, or missing in action from the war.
BYU.edu, Text of the Treaty of Portsmouth:.
Flot.com, Russian Navy history of war.
Frontiers.loc.gov, Russo-Japanese Relations in the Far East. Meeting of Frontiers (Library of Congress)
CSmonitor.com, Treaty of Portsmouth now seen as global turning point from the Christian Science Monitor, by Robert Marquand, 30 December 2005.
Stanford.edu, Lyrics, translation and melody of the song "On the hills of Manchuria" (Na sopkah Manchzhurii).
Google Map with battles of Russo-Japanese War and other important events.
See more Russo-Japanese War Maps at the Persuasive Cartography, The PJ Mode Collection, Cornell University Library
1904 in Japan
1904 in the Russian Empire
1905 in Japan
1905 in the Russian Empire
Conflicts in 1904
Conflicts in 1905
History of Korea
History of Manchuria
Wars involving Japan
Wars involving Russia
Wars involving the Russian Empire
Japan–Russia military relations |
null | null | Everyman (play) | eng_Latn | The of Everyman (The Summoning of Everyman), usually referred to simply as Everyman, is a late 15th-century morality play. Like John Bunyan's 1678 Christian novel The Pilgrim's Progress, Everyman uses allegorical characters to examine the question of Christian salvation and what Man must do to attain it.
Summary
The will is that the good and evil deeds of one's life will be tallied by God after death, as in a ledger book. The play is the allegorical accounting of the life of Everyman, who represents all mankind. In the course of the action, Everyman tries to convince other characters to accompany him in the hope of improving his life. All the characters are also mystical; the conflict between good and evil is shown by the interactions between the characters. Everyman is being singled out because it is difficult for him to find characters to accompany him on his pilgrimage. Everyman eventually realizes through this pilgrimage that he is essentially alone, despite all the personified characters that were supposed necessities and friends to him. Everyman learns that when you are brought to death and placed before God, all you are left with are your own good deeds.
Sources
The play was written in Middle English during the Tudor period, but the identity of the author is unknown. Although the play was apparently produced with some frequency in the seventy-five years following its composition, no production records survive.
There is a similar Dutch-language morality play of the same period called Elckerlijc. In the early 20th century, scholars did not agree on which of these plays was the original, or even on their relation to a later Latin work named Homulus. By the 1980s, Arthur Cawley went so far as to say that the "evidence for … Elckerlijk is certainly very strong", and now Davidson, Walsh, and Broos hold that "more than a century of scholarly discussion has ... convincingly shown that Everyman is a translation and adaptation from the Dutch Elckerlijc".
Setting
The cultural setting is based on the Roman Catholicism of the era. Everyman attains afterlife in heaven by means of good works and the Catholic Sacraments, in particular Confession, Penance, Unction, Viaticum and receiving the Eucharist.
Synopsis
The oldest surviving example of the script begins with this paragraph on the frontispiece:
{| class=wikitable
|Here begynneth a treatyſe how þe hye Fader of Heuen ſendeth Dethe to ſomon euery creature to come and gyue a counte of theyr lyues in this worlde, and is in maner of a morall playe.
|Here begins a treatise how the high Father of Heaven sends Death to summon every creature to come and give account of their lives in this world, and is in the manner of a moral play.
|}
After a brief prologue asking the audience to listen, God speaks, lamenting that humans have become too absorbed in material wealth and riches to follow Him, so He commands Death to go to Everyman and summon him to heaven to make his reckoning. Death arrives at Everyman's side to tell him it is time to die and face judgment. Upon hearing this, Everyman is distressed, so begs for more time. Death denies this, but will allow Everyman to find a companion for his journey.
Everyman's friend Fellowship promises to go anywhere with him, but when he hears of the true nature of Everyman's journey, he refuses to go. Everyman then calls on Kindred and Cousin and asks them to go with him, but they both refuse. In particular, Cousin explains a fundamental reason why no people will accompany Everyman: they have their own accounts to write as well. Afterwards, Everyman asks Goods, who will not come: God's judgment will be severe because of the selfishness implied in Goods's presence.
Everyman then turns to Good Deeds, who says she would go with him, but she is too weak as Everyman has not loved her in his life. Good Deeds summons her sister Knowledge to accompany them, and together they go to see Confession. In the presence of Confession, Everyman begs God for forgiveness and repents his sins, punishing himself with a scourge. After his scourging, Everyman is absolved of his sins, and as a result, Good Deeds becomes strong enough to accompany Everyman on his journey with Death.
Good Deeds then summons Beauty, Strength, Discretion and Five Wits to join them, and they agree to accompany Everyman as he goes to a priest to take sacrament. After the sacrament, Everyman tells them where his journey ends, and again they all abandon him – except for Good Deeds. Even Knowledge cannot accompany him after he leaves his physical body, but will stay with him until the time of death.
Content at last, Everyman climbs into his grave with Good Deeds at his side and dies, after which they ascend together into heaven, where they are welcomed by an Angel. The play closes as the Doctor enters and explains that in the end, a man will only have his Good Deeds to accompany him beyond the grave.
Adaptations
A modern stage production of Everyman did not appear until July 1901 when The Elizabethan Stage Society of William Poel gave three outdoor performances at the Charterhouse in London. Poel then partnered with British actor Ben Greet to produce the play throughout Britain, with runs on the American Broadway stage from 1902 to 1918, and concurrent tours throughout North America. These productions differed from past performances in that women were cast in the title role, rather than men. Film adaptations of the 1901 version of the play appeared in 1913 and 1914, with the 1913 film being presented with an early color two-process pioneered by Kinemacolor.
A version was filmed for Australian TV in 1964.
Another well-known version of the play is Jedermann by the Austrian playwright Hugo von Hofmannsthal, which has been performed annually at the Salzburg Festival since 1920, and adapted into film several times. Frederick Franck published a modernised version of the tale entitled "Everyone", drawing on Buddhist influence. A direct-to-video film of Everyman was made in 2002, directed by John Farrell, which updated the setting to the early 21st century, including Death as a businessman in dark glasses with a briefcase, and Goods being played by a talking personal computer.
A modernized adaptation by Carol Ann Duffy, the Poet Laureate of the United Kingdom, with Chiwetel Ejiofor in the title role, was performed at the National Theatre from April to July 2015.
In December 2016, Moravian College presented Everyman on Trial, a contemporary adaptation written and directed by Christopher Shorr. Branden Jacobs-Jenkins' adaptation, titled Everybody, premiered in 2017 at the Pershing Square Signature Center in New York City.
Notes
References
Schreiber, Earl G. (1975), Everyman in America, Comparative Drama 9.2, pp. 99–115.
Speaight, Robert (1954), William Poel and the Elizabethan revival, London: Heinemann, pp. 161–168.
Editions
Further reading
External links
Full Text, Modern English version of Everyman
A Student Guide to Everyman
1495 plays
1520s plays
British plays adapted into films
English plays
Everyman
Christian plays
Middle English literature
Religious vernacular drama
Works of unknown authorship |
null | null | Redfish | eng_Latn | Redfish is a common name for several species of fish. It is most commonly applied to certain deep-sea rockfish in the genus Sebastes, red drum from the genus Sciaenops or the reef dwelling snappers in the genus Lutjanus. It is also applied to the slimeheads or roughies (family Trachichthyidae), and the alfonsinos (Berycidae).
References
Fish common names |
null | null | Bishop (chess) | eng_Latn | The bishop (♗, ♝) is a piece in the game of chess. It moves and captures along diagonals without jumping over intervening pieces. Each player begins the game with two bishops. One starts between the and the king, the other between the and the queen. The starting squares are c1 and f1 for White's bishops, and c8 and f8 for Black's bishops.
Placement and movement
The bishop has no restrictions in distance for each move but is limited to movement. Bishops, like all other pieces except the knight, cannot jump over other pieces. A bishop captures by occupying the square on which an enemy piece sits.
The bishops may be differentiated according to which wing they begin on, e.g. the king's bishop and queen's bishop. As a consequence of its diagonal movement, each bishop always remains on either the white or black squares, and so it is also common to refer to them as light-squared or dark-squared bishops.
Comparison - other pieces
Versus rook
A rook is generally worth about two pawns more than a bishop (see Chess piece relative value and the exchange). The bishop has access to only half of the squares on the board, whereas all squares of the board are accessible to the rook. On an empty board, a rook always attacks fourteen squares, whereas a bishop attacks no more than thirteen and sometimes as few as seven, depending on how near it is to the center. Also, a king and rook can force checkmate against a lone king, while a king and bishop cannot. However, a king and two bishops on opposite-colored squares can force mate.
Versus knight
Both knights and bishops are worth about 3 pawns. This means bishops are approximately equal in strength to knights, but depending on the game situation either may have a distinct advantage.
Less experienced players tend to underrate the bishop compared to the knight because the knight can reach all squares and is more adept at forking. More experienced players understand the power of the bishop.
Bishops usually gain in relative strength towards the endgame as more pieces are captured and more open lines become available on which they can operate. A bishop can easily influence both wings simultaneously, whereas a knight is less capable of doing so. In an open endgame, a pair of bishops is decidedly superior to either a bishop and a knight, or two knights. A player possessing a pair of bishops has a strategic weapon in the form of a long-term threat to trade down to an advantageous endgame.
Two bishops on opposite-colored squares and king can force checkmate against a lone king, whereas two knights cannot. A bishop and knight can force mate, but with far greater difficulty than two bishops.
In certain positions a bishop can by itself lose a move (see triangulation and tempo), while a knight can never do so. The bishop is capable of skewering or pinning a piece, while the knight can do neither. A bishop can in some situations hinder a knight from moving. In these situations, the bishop is said to be "dominating" the knight.
On the other hand, in the opening and middlegame a bishop may be hemmed in by pawns of both players, and thus be inferior to a knight which can jump over them. A knight check cannot be blocked but a bishop check can. Furthermore, on a crowded board a knight has many tactical opportunities to fork two enemy pieces. A bishop can fork, but opportunities are more rare. One such example occurs in the position illustrated, which arises from the Ruy Lopez: 1.e4 e5 2.Nf3 Nc6 3.Bb5 a6 4.Ba4 Nf6 5.0-0 b5 6.Bb3 Be7 7.d4 d6 8.c3 Bg4 9.h3 Bxf3 10.Qxf3 exd4 11.Qg3 g6 12.Bh6
Game use
Good bishop and bad bishop
In the middlegame, a player with only one bishop should generally place friendly pawns on squares of the color that the bishop cannot move to. This allows the player to control squares of both colors, allows the bishop to move freely among the pawns, and helps fix enemy pawns on squares on which they can be attacked by the bishop. Such a bishop is often referred to as a "good" bishop.
Conversely, a bishop which is impeded by friendly pawns is often referred to as a "bad bishop" (or sometimes, disparagingly, a "tall pawn"). The black light-squared bishop in the French Defense is a notorious example of this concept. However, a "bad" bishop need not always be a weakness, especially if it is outside its own pawn chains. In addition, having a "bad" bishop may be advantageous in an opposite-colored bishops endgame. Even if the bad bishop is passively placed, it may serve a useful defensive function; a well-known quip from GM Mihai Suba is that "Bad bishops protect good pawns."
In the position from the game Krasenkow versus Zvjaginsev, a thicket of black pawns hems in Black's bishop on c8, so Black is effectively playing with one piece fewer than White. Although the black pawns also obstruct the white bishop on e2, it has many more attacking possibilities, and thus is a good bishop vis-à-vis Black's bad bishop. Black resigned after another ten moves.
Fianchetto
A bishop may be fianchettoed, for example after moving the g2 pawn to g3 and the bishop on f1 to g2. This can form a strong defense for the castled king on g1 and the bishop can often exert strong pressure on the long diagonal (here h1–a8). A fianchettoed bishop should generally not be given up lightly, since the resulting in the pawn formation may prove to be serious weaknesses, particularly if the king has castled on that side of the board.
There are nonetheless some modern opening lines where a fianchettoed bishop is given up for a knight in order to double the opponent's pawns, for example 1.d4 g6 2.c4 Bg7 3.Nc3 c5 4.d5 Bxc3+!? 5.bxc3 f5, a sharp line originated by Roman Dzindzichashvili. Giving up a fianchettoed queen's bishop for a knight is usually less problematic. For example, in Karpov–Browne, San Antonio 1972, after 1.c4 c5 2.b3 Nf6 3.Bb2 g6?!, Karpov gave up his fianchettoed bishop with 4.Bxf6! exf6 5.Nc3, doubling Black's pawns and giving him a hole on d5.
Endgame
An endgame in which each player has only one bishop, one controlling the dark squares and the other the light, will often result in a draw even if one player has a pawn or sometimes two more than the other. The players tend to gain control of squares of opposite colors, and a deadlock results. In endgames with same-colored bishops, however, even a positional advantage may be enough to win .
Bishops on opposite colors
Endgames in which each player has only one bishop (and no other pieces besides the king) and the bishops are on opposite colors are often drawn, even when one side has an extra pawn or two. Many of these positions would be a win if the bishops were on the same color.
The position from Wolf versus Leonhardt (see diagram), shows an important defensive setup. Black can make no progress, since the white bishop ties the black king to defending the pawn on g4 and it also prevents the advance ...f3+ because it would simply the pawn – then either the other pawn is exchanged for the bishop (an immediate draw) or the pawn advances (an easily drawn position). Otherwise the bishop alternates between the squares d1 and e2 .
If two pawns are connected, they normally win if they reach their sixth , otherwise the game may be a draw (as above). If two pawns are separated by one they usually draw, but win if they are farther apart .
In some cases with more pawns on the board, it is actually advantageous to have the bishops on opposite colors if one side has weak pawns. In the 1925 game of Efim Bogoljubov versus Max Blümich, (see diagram) White wins because of the bishops being on opposite colors making Black weak on the black squares, the weakness of Black's isolated pawns on the , and the weak doubled pawns on the . The game continued:
29.Kd2 Ke7 30.Kc3 f6 31.Kd4 Be6 32.Kc5 Kd7 33.Kb6 g5 34.Kxa6 Kc7 35.Bb6+ Kc8 36.Bc5 Kc7 37.Bf8 f5 38.Bxg7 f4 39.Bf6 f3 40.gxf3 exf3 41.Bxg5 Bxh3 42.Bf4+
Wrong bishop
In an endgame with a bishop, in some cases the bishop is the "wrong bishop", meaning that it is on the wrong color of square for some purpose (usually promoting a pawn). For example, with just a bishop and a rook pawn, if the bishop cannot control the promotion square of the pawn, it is said to be the "wrong bishop" or the pawn is said to be the wrong rook pawn. This results in some positions being drawn (by setting up a fortress) which otherwise would be won.
History
The bishop's predecessor in medieval chess, shatranj (originally chaturanga), was the alfil, meaning "elephant", which could leap two squares along any diagonal, and could jump over an intervening piece. As a consequence, each fil was restricted to eight squares, and no fil could attack another. The modern bishop first appeared shortly after 1200 in Courier chess. A piece with this move, called a or crocodile, is part of the Grande Acedrex in the compiled in 1283 for King Alfonso X of Castile. The game is attributed to "India", then a very vague term. About half a century later Muḥammad ibn Maḥmud al-Āmulī, in his Treasury of the Sciences, describes an expanded form of chess with two pieces moving "like the rook but obliquely". The bishop was also independently invented in Japan at about the same time (the 13th century), where it formed part of sho shogi and dai shogi; it remains present in modern shogi as the direct descendant of sho shogi.
Derivatives of alfil survive in the languages of the two countries where chess was first introduced within Western Europe—Italian () and Spanish (). It was known as the in French, or the aufin, alphin, or archer in early English.
The earliest references to bishops on the chessboard are two 13th-century Latin texts, and . The etymology of "bishop" comes from Old English bisceop "bishop, high priest," from Late Latin episcopus, from Greek episkopos "watcher, overseer." The term "bishop" as applied specifically to the chess piece was first recorded in the 16th century, with the first known written example dating back to 1560s. In all other Germanic languages, except for Icelandic, it is called various names, all of which directly translate to English as "runner" or "messenger". In Icelandic, however, it is called "", with the same meaning as in English. The use of the term in Icelandic predates that of the English language, as the first mentioning of "" in Icelandic texts dates back to the early part of the 14th century, while the 12th-century Lewis Chessmen portray the bishop as an unambiguously ecclesiastical figure. In the Saga of Earl Mágus, which was written in Iceland somewhere between 1300–1325, it is described how an emperor was checkmated by a bishop. This has led to some speculations as to the origin of the English use of the term "bishop".
The canonical chessmen date back to the Staunton chess set of 1849. The piece's deep groove symbolizes a bishop's (or abbot's) mitre. Some have written that the groove originated from the original form of the piece, an elephant with the groove representing the elephant's tusks. The English apparently chose to call the piece a bishop because the projections at the top resembled a mitre. This groove was interpreted differently in different countries as the game moved to Europe; in France, for example, the groove was taken to be a jester's cap, hence in France the bishop is called "", the "jester" and in Romania the (madman).
In some Slavic languages (e.g. Czech/Slovak) the bishop is called střelec/strelec, which directly translates to English as a "shooter" meaning an archer, while in others it is still known as "elephant" (e.g. Russian ). In South Slavic languages it is usually known as lovac, meaning "hunter", or laufer, taken from the German name for the same piece (laufer is also a co-official Polish name for the piece alongside goniec). An alternative name for bishop in Russian is officer (); it is also called () in Greek, () in Belarusian and oficeri in Albanian.
In Mongolian and several Indian languages it is called the camel.
In Lithuanian it is the , a kind of military commander in medieval Lithuania.
Name translations
Unicode
Unicode defines two codepoints for bishop:
♗ U+2657 White Chess Bishop (HTML ♗)
♝ U+265D Black Chess Bishop (HTML ♝)
See also
Notes
References
External links
Piececlopedia: Bishop by Fergus Duniho and Hans Bodlaender, The Chess Variant Pages
Chess pieces |
null | null | British Sign Language | eng_Latn | British Sign Language (BSL) is a sign language used in the United Kingdom (UK), and is the first or preferred language among the Deaf community in the UK. Based on the percentage of people who reported 'using British Sign Language at home' on the 2011 Scottish Census, the British Deaf Association estimates there are 151,000 BSL users in the UK, of which 87,000 are Deaf. By contrast, in the 2011 England and Wales Census 15,000 people living in England and Wales reported themselves using BSL as their main language. People who are not deaf may also use BSL, as hearing relatives of deaf people, sign language interpreters or as a result of other contact with the British Deaf community. The language makes use of space and involves movement of the hands, body, face, and head.
History
The Beginning
BSL is the creation of the British Deaf community, who have experienced discrimination over many centuries. All sign languages have origins in gestural communication developed between deaf children and hearing adults. Unlike home sign, which does not pass between generations, sign languages are shared by a large community of signers.
Records show the existence of a sign language within deaf communities in England as far back as the 15th century. The History of the Syon Monastery at Lisbon and Brentford, published in 1450, contains descriptions of signs - some of which are still in use. The earliest documented use of sign language is the registry records of a marriage ceremony between Thomas Tilsye and Ursula Russel in 1576.
Richard Carew's Survey of Cornwall (1602) includes a vivid description of Edward Bone, a deaf servant, meeting his deaf friend Kempe. Bone had some knowledge of English and was able to lipread, but appeared to prefer signing. Carew described the situation thus:Somewhat neerre the place of his [Bone's] birth, there dwelt another, so affected, or rather defected, whose name was Kempe: which two, when they chaunced to meete, would use such kinde embracements, such stranfe, often, and earnest tokenings, and such heartie laughtes, and other passionate gestures, that their want of a tongue, seemed rather an hinderance to other conveiving [understanding] them, then to their conceiving one another.John Bulwer, who had an adopted deaf daughter Chirothea Johnson, authored four late-Renaissance texts related to deafness, sign language and the human body: Chirologia (1644), Philocopus (1648), Pathomyotamia (1649) and Anthropometamorphosis (1650). In particular, Chirologia focuses on the meanings of gestures, expressions and body language, and describes signs and gestures in use at the time, some of which resemble signs still in use, while Philocopus explores the use of lipreading by deaf people and the possibility of deaf education, and is dedicated to Bulwar's two deaf brothers.
Another writer of the same time, George Dalgarno, recognised that sign language was unrelated to English. In 1661 he wrote that "The deaf man has no teacher at all and through necessity may put him upon... using signs, yet those have no affinity to the language by which they that are about him do converse among themselves."
Finally, the diarist Samuel Pepys described a conversation between George Downing and a deaf boy in November 1666:But, above all, there comes in the dumb boy that I knew in Oliver's time, who is mightily acquainted here, and with Downing; and he made strange signs of the fire, and how the King was abroad, and many things they understood, but I could not...British Sign Language has evolved, as all languages do, from these origins by modification, invention and importation.
Early Deaf Education
The Braidwood schools
Thomas Braidwood, a teacher from Edinburgh, founded 'Braidwood's Academy for the Deaf and Dumb' in 1760 which is believed to be the first school for deaf children in Britain. The school primarily taught oral communication methods, as described by Francis Green - whose son attended the Braidwood school - in the anonymous treatise Vox oculis subjecta. In this account, Green describes how his son Charles would surely develop "a perfect acquaintance with language both oral and written", and how deaf pupils were given "a tolerable general understanding of their own language [English] so as to read, write, and speak it, with ease". Green also describes Braidwood's views of spoken language:Mr Braidwood hath frequently intimated to me, as an opinion founded upon his experience in this art, that articulate or spoken language hath so great and essential a tendency to confirm and enlarge ideas, above the power of written language, that it is almost impossible for deaf persons, without the use of speech, to be perfect in their ideas.Joseph Watson was trained as a teacher of the deaf under Thomas Braidwood. He eventually left in 1792 to become the headmaster of the Asylum for the Deaf and Dumb in Bermondsey. He described his teaching methods in detail in his book, On the Education of the Deaf and Dumb (1809), where he opposed the use of signed versions of spoken language such as the Signed French used in the Paris school. The book contains lists of vocabulary and plates designed to encourage a child to acquire an understanding of written and spoken language.
International links
Although the Braidwood school focused on speech, it also used an early form of sign language, the combined system, which was the first codification of British Sign Language. The Braidwood school later moved to London and was visited by Abbé Sicard and Laurent Clerc in 1815, at the same time that an American Protestant minister, Thomas Hopkins Gallaudet, travelled to Europe to research teaching of the deaf.
André-Daniel Laffon de Ladebat, one of the French visitors to the Braidwood school, provided a vivid description of Laurent Clerc's meeting with the deaf children in the bilingual English/French book, A collection of the Most Remarkable Definitions and Answers of Massieu and Clerc, Deaf and Dumb. Laurent Clerc, who was deaf, was overjoyed to find fellow sign language users:As soon as Clerc beheld this sight [of the children at dinner] his face became animated; he was as agitated as a traveller of sensibility would be on meeting all of a sudden in distant regions, a colony of his own countrymen... Clerc approached them. He made signs and they answered him by signs. The unexpected communication cause a most delicious sensation in them and for us was a scene of expression and sensibility that gave us the most heart-felt satisfaction.The Braidwood schools refused to teach Gallaudet their methods. Gallaudet then travelled to Paris and learned the educational methods of the French Royal Institution for the Deaf, a combination of Old French Sign Language and the signs developed by Abbé de l’Épée. As a consequence American Sign Language today has a 60% similarity to modern French Sign Language and is almost unintelligible to users of British Sign Language. Gallaudet went on to establish the American School for the Deaf in 1817, which focused on manual communication and ASL, in contrast to the oral methods used in the UK.
Late 19th - 20th Century
Until the 1940s sign language skills were passed on unofficially between deaf people often living in residential institutions. Signing was actively discouraged in schools by punishment and the emphasis in education was on forcing deaf children to learn to lip read and finger spell. From the 1970s there has been an increasing tolerance and instruction in BSL in schools. The language continues to evolve as older signs such as alms and pawnbroker have fallen out of use and new signs such as internet and laser have been coined. The evolution of the language and its changing level of acceptance means that older users tend to rely on finger spelling while younger ones make use of a wider range of signs.
BSL users campaigned to have BSL recognised on an official level. Police and Criminal Evidence Act 1984 mandates the provision of interpreters. On 18 March 2003 the UK government formally recognised that BSL is a language in its own right. In 2022, Rosie Cooper introduced the British Sign Language Bill to recognise BSL as an official language, which was backed by the government.
Linguistics
Linguistics is the study of language, including those like BSL that are not carried by sound. In all sign languages the great majority of 'words' (hand gestures) cannot be understood in other sign languages. How one language signs a certain number would be different than how another language signs it. The way sentences are constructed (syntax) differs from sign language to sign language, just as with different spoken languages. British Sign Language is described as a 'spatial language' as it "moves signs in space."
Phonology
Like many other sign languages, BSL phonology is defined by elements such as handshape, orientation, location, movement, and non-manual features. There are phonological components to sign language that have no meaning alone but work together to create a meaning of a signed word: hand shape, movement, location, orientation and facial expression. The meanings of words differ if one of these components is changed. Signs can be identical in certain components but different in others, giving each a different meaning. Facial expression falls under the 'non-manual features' component of phonology. These include "eyebrow height, eye gaze, mouthing, head movement, and torso rotation."
Grammar
In common with other languages, whether spoken or signed, BSL has its own grammar which govern how phrases are signed. BSL has a particular syntax. One important component of BSL is its use of proforms. A proform is “...any form that stands in the place of, or does the job of, some other form." Sentences are composed of two parts, in order: the subject and the predicate. The subject is the topic of the sentence, while the predicate is the commentary about the subject.
BSL uses a topic–comment structure. Topic-comment means that the topic of the signed conversation is first established, followed by an elaboration of the topic, being the 'comment' component. The canonical word order outside of the topic–comment structure is object–subject–verb (OSV), and noun phrases are head-initial.
Relationships with other sign languages
Although the United Kingdom and the United States share English as the predominant oral language, British Sign Language is quite distinct from American Sign Language (ASL) - having only 31% signs identical, or 44% cognate. BSL is also distinct from Irish Sign Language (ISL) (ISG in the ISO system) which is more closely related to French Sign Language (LSF) and ASL.
It is also distinct from Signed English, a manually coded method expressed to represent the English language.
The sign languages used in Australia and New Zealand, Auslan and New Zealand Sign Language respectively, evolved largely from 19th century BSL, and all retain the same manual alphabet and grammar and possess similar lexicons. These three languages may technically be considered dialects of a single language (BANZSL) due to their use of the same grammar and manual alphabet and the high degree of lexical sharing (overlap of signs). The term BANZSL was coined by Trevor Johnston and Adam Schembri.
In Australia deaf schools were established by educated deaf people from London, Edinburgh and Dublin. This introduced the London and Edinburgh dialects of BSL to Melbourne and Sydney respectively and Irish Sign Language to Sydney in Roman Catholic schools for the deaf. The language contact post secondary education between Australian ISL users and 'Australian BSL' users accounts for some of the dialectal differences we see between modern BSL and Auslan. Tertiary education in the US for some deaf Australian adults also accounts for some ASL borrowings found in modern Auslan.
Auslan, BSL and NZSL have 82% of signs identical (using concepts from a Swadesh list). When considering similar or related signs as well as identical, they are 98% cognate. Further information will be available after completion of the BSL corpus, allows for comparison with the Auslan corpus, and the New Zealand Sign Language project. There continues to be language contact between BSL, Auslan and NZSL through migration (deaf people and interpreters), the media (television programmes such as See Hear, Switch, Rush and SignPost are often recorded and shared informally in all three countries) and conferences (the World Federation of the Deaf Conference – WFD – in Brisbane 1999 saw many British deaf people travelling to Australia).
Makaton, a communication system for people with cognitive impairments or other communication difficulties, was originally developed with signs borrowed from British Sign Language. The sign language used in Sri Lanka is also closely related to BSL despite the oral language not being English, demonstrating the distance between sign languages and spoken ones.
Usage
BSL has many regional dialects. Certain signs used in Scotland, for example, may not be understood immediately, or not understood at all, by those in Southern England, or vice versa. Some signs are even more local, occurring only in certain towns or cities (such as the Manchester system of number signs). Likewise, some may go in or out of fashion, or evolve over time, just as terms in oral languages do. Families may have signs unique to them to accommodate for certain situations or to describe an object that may otherwise require fingerspelling.
Many British television channels broadcast programmes with in-vision signing, using BSL, as well as specially made programmes aimed mainly at deaf people such as the BBC's See Hear and Channel 4's VEE-TV.
BBC News broadcasts in-vision signing at 07:00-07:45, 08:00-08:20 and 13:00-13:45 GMT/BST each weekday. BBC Two also broadcasts in-vision signed repeats of the channel's primetime programmes between 00:00 and 02:00 each weekday. All BBC channels (excluding BBC One, BBC Alba and BBC Parliament) provide in-vision signing for some of their programmes. In 2020, 5.5% of Channel 4's programming was signed, including popular shows such as Hollyoaks and Gogglebox.
BSL is used in some educational establishments, but is not always the policy for deaf children in some local authority areas. The Let's Sign BSL and fingerspelling graphics are being developed for use in education by deaf educators and tutors and include many of the regional signs referred to above.
In Northern Ireland, there are about 4,500 users of BSL and 1,500 users of Irish Sign Language, an unrelated sign language. A hybrid version, dubbed "Northern Ireland Sign Language" is also used.
In 2019, over 100 signs for scientific terms, including 'deoxyribonucleotide' and 'deoxyribonucleoside', were added to BSL, after being conceived by Liam Mcmulkin, a deaf graduate of the University of Dundee, who had found finger-spelling such words tiresome, during his degree course.
Number of BSL users
In 2016 the British Deaf Association says that, based on official statistics, it believes there are 151,000 people who use BSL in the UK, and 87,000 of these are deaf. This figure does not include professional BSL users, interpreters, translators, etc. unless they use BSL at home.
British Sign Language Dictionary
The dictionary was compiled for the BDA by the Deaf Studies Research Unit at the University of Durham. It depicts over 1,800 signs through pictures and diagrams, each sign accompanied by definitions, explanations and usage. The signs are ordered not according to the alphabetical order of the English translation, but rather according to the phonological characteristics of the language. For example, signs that are based on the "fist" handshape come before signs based on the "open hand" handshape.
The dictionary was edited by David Brien, assisted by a team composed by Mary Brennan, Clark Denmark, Frances Elton, Liz Scoot Gibson, Graham Turner and Dorothy Miles, among others.
The Dictionary was published in 1992. The foreword was written by Princess Diana, who was the patron of the BDA.
Learning British Sign Language
British Sign Language can be learnt throughout the UK and three examination systems exist. Courses are provided by community colleges, local centres for deaf people and private organisations. A teaching qualification program was started by the British Deaf Association (BDA) in 1984 at the University of Durham, called BSL Tutor Training Course, which closed in 1999.
National awarding organisations run training for BSL teachers. Each of these organisations have their own curricula, teaching materials and resources.
Becoming a BSL / English interpreter
There are two qualification routes: via post-graduate studies, or via National Vocational Qualifications. Deaf Studies undergraduate courses with specific streams for sign language interpreting exist at several British universities; post-graduate level interpreting diplomas are also on offer from universities and one private company. Course entry requirements vary from no previous knowledge of BSL to NVQ level 6 BSL (or equivalent).
The qualification process allows interpreters to register with the National Registers of Communication Professionals with Deaf and Deafblind People (NRCPD), a voluntary regulator. Registrants are asked to self-certify that they have both cleared a DBS (Disclosure and Barring Service) check and are covered by professional indemnity insurance. Completing a level 3 BSL language assessment and enrolling on an approved interpreting course allows applications to register as a TSLI (Trainee Sign Language Interpreter). After completing an approved interpreting course, trainees can then apply to achieve RSLI (Registered Sign Language Interpreter) status. RSLIs are currently required by NRCPD to log Continuous Professional Development activities. Post-qualification, specialist training is still considered necessary to work in specific critical domains.
Communication Support Workers
Communication Support Workers (CSWs) are professionals who support the communication of deaf students in education at all ages, and deaf people in many areas of work, using British Sign Language and other communication methods such as Sign Supported English. The qualifications and experience of CSWs varies: some are fully qualified interpreters, others are not.
Let Sign Shine
Let Sign Shine is a campaign started by Norfolk teenager Jade Chapman to raise the awareness of British Sign Language (BSL) and attract signatures for a petition for BSL to be taught in schools. The campaign's petition to the Parliament of the United Kingdom has attracted support from over four thousand people.
Chapman was nominated for the Bernard Matthews Youth Award 2014 for her work and devotion to raising awareness of the importance of sign language. Chapman won the education award category and was presented with an award by Olympic swimmer Rebecca Adlington.
Chapman was also awarded an Outstanding Achievement Award from the Radio Norwich 99.9 Local Hero Awards on 7 October 2015. The award ceremony featured a performance by Alesha Dixon.
Having been donated £1,000 from the Bernard Matthews Youth Award, Let Sign Shine used this to start a British Sign Language course at Dereham Neatherd High School.
See also
Languages in the United Kingdom
Makaton
References
External links
British Deaf Association Sign Language Resource
BSL Sign Language Dictionary
BSL SignBank
BSL Online Dictionary
Information and Communication Technology Dictionary
Languages of the United Kingdom
BANZSL Sign Language family
Deaf culture in the United Kingdom
Languages of Scotland
Languages of Wales
Languages of England
Sign languages of the United Kingdom |
null | null | Mahātmā | eng_Latn | Mahātmā () is a honorific that in Sanskrit means "great soul" (महात्मा mahātmā: महा mahā 'great' + आत्मा ātmā 'soul').
The term is commonly used for Mohandas Karamchand Gandhi, who is often referred to simply as "Mahatma Gandhi". Albeit less frequently, this epithet is also been used with regard to such people as Basava (1131–1167), Swami Shraddhananda (1856–1926), Lalon Shah (1772–1890), Ayyankali (1863–1941), and Jyotirao Phule (1827–1890).
The term mahātmā has also been historically used for a class of religious scholars in Jainism; for the selected religious leaders in Theosophy; and for local religious teachers in the Divine Light Mission church.
In Theosophy
The word, used in a technical sense, was popularized in theosophical literature in the late 19th century, when Helena Blavatsky, one of the founders of the Theosophical Society, claimed that her teachers were adepts (or Mahatmas) who reside in Asia. According to the Theosophical teachings, the Mahatmas are not disembodied beings, but highly evolved people involved in overseeing the spiritual growth of individuals and the development of civilizations. Blavatsky was the first person in modern times to claim contact with these Adepts, especially the "Masters" Koot Hoomi and Morya. Alvin Boyd Kuhn wrote about mahātmās:
The Masters whom Theosophy presents to us are simply high-ranking students in life's school of experience. They are members of our own evolutionary group, not visitants from the celestial spheres. They are supermen only in that they have attained knowledge of the laws of life and mastery over its forces with which we are still struggling.
After Blavatsky's death in 1891, numerous individuals have claimed to be in contact with her Adept Teachers. These individuals have stated that they are new "messengers" of the Masters and they have conveyed various esoteric teachings. Currently, various Theosophical movements refer to them as Ascended Masters, although their character and teachings are in several respects different from those described by Blavatsky.
Criticism
There has been a great deal of controversy concerning the existence of adepts. Blavatsky's critics have doubted the existence of her Masters. See, for example, W. E. Coleman's "exposes".
K. Paul Johnson in his books speculates that the "Masters" that Blavatsky wrote about and produced letters from were actually idealizations of people who were her mentors. Aryel Sanat, author of The inner life of Krishnamurti: private passion and perennial wisdom, wrote that Johnson "claims in all of his books that there were no Masters at all in early Theosophical Society history, and that Helena Blavatsky invented them (as others had claimed she had invented her travels)." Sanat wrote that Johnson "deliberately ignores the main sources of evidence for their real physical existence." What Sanat thought these were is not made quite clear.
Mohandas Karamchand Gandhi as BAPU Mahātmā
According to some authors, Rabindranath Tagore is said to have used this title for Gandhi on 6 March 1915. Some claim that he was called Mahatma by the residents of Gurukul Kangadi in April 1915, and he in turn called the founder Munshiram a Mahatma (who later became Swami Shraddhananda). However, a document honoring him with the title "Mahatma" on 21 January 1915, in Jetpur, Gujarat, by Nautamlal Bhagvanji Mehta is preserved at the National Gandhi Museum in New Delhi, India. A private letter from Gandhi's friend Pranjivan Mehta to Gopal Krishna Gokhale in 1909 referred to Gandhi with the title and thus remains the earliest written record available to use the title. The use of the term Mahatma in Jainism to denote a class of lay priests, has been noted since the 17th century. A Mahatma is someone who practices Trikaranasuddhi.
In the Divine Light Mission
The Divine Light Mission (DLM) was a Sant Mat-based movement begun in India in the 1930s by Hans Ji Maharaj and formally incorporated in 1960. The DLM had as many as 2,000 Mahatmas, all from India or Tibet, who taught the DLM's secret meditation techniques called "Knowledge". The Mahatmas, called "realized souls", or "apostles", also served as local leaders. After Hans Ji's death in 1966 his youngest son, Prem Rawat (known then as Guru Maharaj Ji or Bagyogeshwar), succeeded him. The young guru appointed some new Mahatmas, including one from the United States. In one incident, a prominent Indian Mahatma nearly beat a man to death in Detroit for throwing a pie at the guru. In the early 1980s, Prem Rawat replaced the Divine Light Mission organization with the Elan Vital and replaced the Mahatmas with initiators. The initiators did not have the revered status of the Mahatmas, and they were drawn mostly from Western followers. In the 2000s, the initiators were replaced by a video in which Rawat teaches the techniques himself.
In Jainism
Among the Jains the term Mahatma is used for the class for scholars who are householders.
Mahatma Hirananda of Mewar
The colophon of the Mewar Ramayana, described as "one of the most beautiful manuscripts in the world", states that the text, commissioned by Acharya Jasvant for the library of Maharana Jagat Singh I of Mewar, was written by one Mahatma Hirananda and finished on Friday 25 November 1650. Being a Jain scribe, Mahatma Hirananda incorporated traditional Jain scribal elements into the manuscript.
Jain Mahatmas in Dabestan-e Mazaheb
The Dabestan-e Mazaheb, written around 1655 CE, is a text that describes various religions and philosophies the author encountered. Its Section 11 is dedicated to Jainism. It states: "Similar to the dervishes of both classes (Monk and Jatis) is a third sect, called Mahá-átma; group separated from the Oswal Jains, have the dress and appearance of Jatis; however, do not pluck their hair with tweezers, but cut it. They live the social life and spread the teaching of Mahavir and other Jain Tirthankars in the society. They accumulate money, cook their meals in their houses, drink cold water, and take to them a wife." The term Mahatma was thus used for priests/scholars who were not celibate.
See also
Ascended master
Initiation (Theosophy)
Theosophy
Alice A.Bailey
Helena Blavatsky
Helena Roerich
Benjamin Creme
References
Bibliography
French, Brendan James (2000) The theosophical masters: an investigation into the conceptual domains of H.P. Blavatsky and C.W. Leadbeater. Ph.D. thesis, University of Sydney, Department of Religious Studies. Sydney.
Johnson, K. Paul (1994) The Masters Revealed: Madam Blavatsky and Myth of the Great White Brotherhood. Albany, New York: State University of New York Press.
Johnson, K. Paul (1995) Initiates of Theosophical Masters. Albany, New York: State University of New York Press.
Kalnitsky, Arnold (2003) The Theosophical Movement of the Nineteenth Century: The Legitimation of the Disputable and the Entrenchment of the Disreputable. University of South Africa. Dissertation: 443 pp.
Kuhn, Alvin Boyd (1930) Theosophy: A Modern Revival of Ancient Wisdom. Ph.D. Thesis. Whitefish, Montana: Kessinger Publishing. .
Further reading
Dutta, Krishna, and Andrew Robinson. Rabindranath Tagore: An Anthology. Picador/Macmillan: London, 1997.
External links
The theosophical Mahatmas and Their Letters
Theosophy
Theosophical philosophical concepts
Titles in India
Mahatma Gandhi |
null | null | Star Wars Jedi Knight II: Jedi Outcast | eng_Latn | Star Wars Jedi Knight II: Jedi Outcast is a 2002 first and third-person shooter video game. The Microsoft Windows and OS X versions were developed by Raven Software, and the Xbox and GameCube versions by Vicarious Visions; most versions were published by LucasArts, with only the OS X version published by Aspyr. The game is a sequel to 1997's Star Wars Jedi Knight: Dark Forces II, and the third main installment in the Star Wars: Jedi Knight series. The single-player campaign, set in the fictional Star Wars expanded universe two years after the Mysteries of the Sith expansion for Dark Forces II, follows returning protagonist Kyle Katarn, a mercenary working for the New Republic and former Jedi who cut his connection to the Force. Katarn must return to his Jedi ways to stop a branch of the Imperial Remnant led by the Dark Jedi Desann from empowering their army with the Force.
Jedi Outcast was developed using a more powerful game engine, id Tech 3. The hack and slash combat introduced in Dark Forces II was heavily reworked, becoming the main focus of the gameplay instead of the shooter elements, which are prominent only during the first few missions of the game. Players may wield blasters, lightsabers, and Force powers to engage enemies, with the latter being recommended in later stages of the game, as numerous lightsaber-wielding enemies are introduced. Jedi Outcast also features a multiplayer mode that allows players to compete in several different game modes online or over a local area network.
Upon release, the game received largely positive reviews from critics, with its story and lightsaber combat being the main praised elements. A sequel and the final installment in the Jedi Knight series, Star Wars Jedi Knight: Jedi Academy, was released in 2003. Both Jedi Outcast and Jedi Academy are considered landmark Star Wars titles for their revolutionary lightsaber combat. In September 2009, the game was re-released onto Steam and Direct2Drive alongside the rest of the Jedi Knight series. A Nintendo Switch and PlayStation 4 port with no multiplayer mode was released in September 2019.
On November 15, 2021, Microsoft announced that in celebration of 20 years of Xbox, they would be adding over 70 more games to their backwards compatibility program. Included in this list was Jedi Outcast, making it availabe to be played on Microsoft's Xbox One and Xbox Series X/S consoles.
Gameplay
Jedi Outcast allows the player to wield a variety of firearms from the Star Wars franchise, as well as lightsabers and Force powers. The player can choose whether to use first or third-person perspective for each weapon, including the lightsaber. Combat is standard for the shooter genre, offering players an array of energy and projectile weapons, plus a variety of explosives. Players have health and shield meters, each of which is replenished separately.
Jedi Outcast places a strong emphasis on lightsaber combat. As in the films, lightsabers can be used to deflect shots from blasters. The game offers three lightsaber styles; fast, medium and strong, with each style differing from the others in terms of the speed of attacks and damage dealt. There are also a number of combos, many of which are unique to the selected saber style.
Force powers (such as Push, Jump, and Lightning) are available in both single-player and multiplayer modes, but more powers can be used in the latter. The use of powers is restricted by a "Force Meter", which depletes with each use and gradually refills over time. The "level" of a Force power determines the strength of that power and the amount depleted from the Force meter during its use. The multiplayer mode divides players into Light Siders and Dark Siders, pitting each side against the other in team battles. Each side has access to both shared "Neutral" Force powers, which are mostly focused on increasing speed and athletic ability. There are also numerous powers unique to both Light and Dark sides. As in the previous games, Light Side powers are mainly focused around protection and healing, while Dark Side powers are openly aggressive. Unlike previous games, however, Kyle does not exclusively select Light or Dark powers in the single-player, instead receiving a selection of both.
Single-player
The single-player campaign follows Kyle Katarn as he moves through the levels in a linear manner, meeting friendly and hostile non-player characters (NPCs). Friendly NPCs will occasionally assist the player in combat. In addition to combat, the campaign features a variety of puzzles.
When the game starts, Kyle has forsaken The Force after the events of the previous game, and as such, the player has no access to a lightsaber or any Force powers. However, after the first two missions, Kyle regains his Force abilities. As the game progresses the number of powers available, and their strength, increase. Progression of Force abilities is fixed, and cannot be customized. Having previously fallen to the Dark Side, Kyle has access to both Light Side powers (such as Force Heal and Jedi Mind Trick) and Dark Side powers (such as Force Lightning and Force Grip), along with neutral ones (such as Force Speed, Force Jump, Force Pull and Force Push).
Multiplayer
Jedi Outcast features a set of multiplayer modes. In the PC and Macintosh versions, these can be played over a LAN or the Internet, but combat is limited to two players on the console versions. Multiplayer works by allowing members of the community to run their own dedicated servers that players can join via the in-game Multiplayer Server List, with server owners being able to dictate custom mods, maps and player models for their server. There are a variety of standardized game modes in Multiplayer (such as "Free-for-All", "Team Deathmatch" and "Capture the flag") which can be played with other players, bots, or both. Server owners can also build their own game modes for their server if they wish to do so. Players have limited customization control over their avatars, however there is the ability for players to import custom models on participating servers. They can choose the player model (with access to nearly every character in the game, as well as some characters from the films not seen in the single-player mode) and lightsaber color. Before a match, the server specifies the game rules, including "Force ranking", which controls how many points the players have available to allocate into Force powers. Players then customize their powers for the match. The server can also choose to disable normal weapons so as to create lightsaber-only matches.
Synopsis
Setting and characters
The single-player story is set in 12 ABY, eight years after the events of Return of the Jedi and around two years after Mysteries of the Sith. As with the previous installments in the series, the player controls Kyle Katarn (voiced by Jeff Bennett), a former Jedi who has cut his links with the Force after almost succumbing to the Dark Side. At the start of the game he is a mercenary working for the New Republic.
Over the course of the game, Kyle is joined by several characters that assist him in varying measures, including his mercenary partner and love interest Jan Ors (Vanessa Marshall); the sophisticated administrator of Cloud City, Lando Calrissian (Billy Dee Williams); and Jedi Grandmaster and leader of the Jedi Academy on Yavin IV, Luke Skywalker (Bob Bergen). The player also receives help from other Jedi and New Republic soldiers, while Mon Mothma (Carolyn Seymour), Chief-of-State of the New Republic, assigns Kyle and Jan missions during the early stages of the game.
The story features four major antagonists and bosses: Desann (Mark Klastorin), a former student of the Jedi Academy, who turned to the Dark Side after killing a fellow student and leaving the Order; Tavion (Kath Soucie), Desann's apprentice; Galak Fyyar (Steve Blum), an admiral in the Imperial Remnant and Desann's second-in-command; and Reelo Baruk (Kevin Michael Richardson), a crime lord with ties to the Imperial Remnant who poses as a "respectable garbage collector" on Nar Shaddaa. Throughout the game, players fight different types of enemies, including mercenaries and Imperial stormtroopers. Halfway through the story, lightsaber-wielding enemies are introduced, namely the Reborn (thugs and soldiers who have been artificially infused with the Force), and Shadowtroopers (who sport lightsaber-resistant cortosis armor that also allows them to briefly turn invisible).
Plot
The game begins with Kyle Katarn and Jan Ors investigating a supposedly abandoned Imperial outpost on Kejim, finding it to be crawling with Imperial forces. They fight their way through the base, discovering a research center studying crystals similar to those used to power lightsabers. Tracing the crystals' origin, Kyle and Jan travel to Artus Prime, a mining colony turned into an Imperial stronghold, where the miners have been enslaved and experimented upon. Katarn thwarts the Imperial operation, but Jan is captured by the Dark Jedi Desann and his apprentice Tavion. Kyle tries to rescue her, but, having forsaken the ways of the Jedi, he is easily defeated by Desann, who orders Tavion to kill Jan before they leave.
An enraged Kyle travels to the Valley of the Jedi (a major plot element in Dark Forces II) to regain his Force powers, and then to the Jedi Academy to reclaim his lightsaber, which he had left with Luke Skywalker. Luke reveals Desann's origins to Kyle, and senses his anger through the Force, so he requests that Kyle complete a series of trials to prove he will not succumb to the Dark Side again. Kyle passes the tests and obtains his lightsaber, before travelling to Nar Shaddaa, where Luke told him to seek out Reelo Baruk, a Rodian gangster. While searching for him, Kyle runs into Lando Calrissian, who has been imprisoned in Reelo's dungeons. Lando reveals that Reelo is part of a cortosis smuggling operation in Cloud City, which has been taken over by the Empire and Reelo's men. After Kyle frees Lando, the pair make their escape in the latter's ship, the Lady Luck; killing Reelo when he tries to stop them.
On Bespin, Lando drops Kyle off at the bottom of Cloud City, and while making his way up the city structure, the latter has his first encounters with the Reborn, Desann's Force-wielding soldiers. After helping Lando's men reclaim control of the city, Kyle finds and defeats Tavion, who, fearing for her life, reveals that Jan is alive and onboard Galak Fyyar's Star Destroyer, the Doomgiver. Jan's faked death was just a ploy to trick Kyle into going to the Valley of the Jedi, which allowed Desann to follow him there and tap its power. Kyle spares Tavion and takes her ship to infiltrate the Doomgiver, which is docked at the Cairn Installation, an Imperial base hidden on an asteroid in the Lenico Belt.
At the base, Kyle runs into Luke, who informs him that Desann has used the Valley's energy to empower an army of Reborn which could number in the thousands. After fighting several Reborn together, Kyle and Luke split up to find the Doomgiver. Along the way, the former has his first encounter with the Shadowtroopers, and discovers that the base is actually an assault ship construction facility, which is preparing for a full-scale planetary assault. As Kyle boards the Doomgiver, he sees Luke fighting Desann outside the ship, but can't intervene because the Star Destroyer jumps into hyperspace. Kyle uses the ship's communications array to contact Rogue Squadron, and finds Jan in the detention block, who reveals that Desann was not specifically interested in the Valley of the Jedi, but rather the Jedi Academy, which he is invading just now to harvest its Force power. After killing Fyyar and destroying the Doomgivers shield reactor, which leaves the ship open for attack, Kyle escapes with Jan in a safe pod, moments before Rogue Squadron destroy the Doomgiver.
Kyle and Jan arrive on Yavin IV, in the middle of the Imperial invasion, and split up: the latter goes to assist the New Republic starfighters, while the former makes his way to the Jedi Academy on foot. After helping the Jedi students and Republic soldiers fend off the assault, Kyle ventures into the Academy's underground maze, and confronts Desann as he tries to absorb the power from the Force nexus at the centre. Kyle informs Desann of the Doomgivers destruction and the Imperial army's defeat, and offers him the chance to rejoin the Jedi, but an enraged Desann refuses and attacks him. Kyle kills Desann and returns to the surface, where he reunites with Jan, as the New Republic arrests the surviving Imperial forces. Later, Luke thanks Kyle and Jan for their assistance, and offers to safeguard the former's lightsaber once more, but Kyle politely refuses, saying he is not ready to forsake the Force again.
Development
On May 17, 2001, at E3 2001, LucasArts announced that Raven Software were developing a third game in the Jedi Knight series. Some plot details were given, such as the locations visited in the game; Cloud City, Yavin IV, Smuggler's Moon and some planets original to the game. The following day at E3, LucasArts gave a demonstration of the game, showing the lightsaber and Force combat as well as the "buddy" system: in which certain NPCs would fight alongside the player. Technical details were also revealed: the game would use id Software's Quake III Arena engine, and the GHOUL 2 animation system, seen in Raven's Soldier of Fortune II: Double Helix. The polygon capacity of the engine had also been doubled. The game was also on display at id's QuakeCon 2001, where enemy AI and combat were demonstrated. LucasArts announced that the game would feature multiplayer, although due to being early in development little information was revealed.
On July 17, 2001, multiplayer developer Pat Lipo made a post on his .plan file, revealing that Rich Whitehouse had been brought onto the development team to handle development of the game's multiplayer bots. Whitehouse moved on to tackle the entirety of the game's multiplayer codebase, and was subsequently credited as the game's sole multiplayer programmer. Designer Chris Foster has stated that most members of the team had many responsibilities due to its small size. One of the main tools used to design levels was Radiant.
On January 16, 2002, LucasArts launched a new website for Jedi Outcast featuring an overview of the game and information regarding characters, weapons and Force powers. An FAQ, screenshots, concept art, images of player models and downloadable wallpapers were also available. A teaser trailer was released on February 8, showing the game's combat, weapons, characters and environments. On March 13, 2002, LucasArts announced that the game was on track for release later in the month. Two days later, they announced Jedi Outcast was ready for release, set to go on sale by March 29. A new trailer was also released. After the release, game programmer Mike Gummelt revealed that a request from the management led the team to disable the ability to sever the heads of non-droid enemies. He has also cited Bushido Blade as the inspiration for a lightsaber combat system which requires players to wait and strike at the right moment.
Release
Jedi Outcast shipped on March 26, 2002.
The game's SDK was released on April 22, 2002. This included a level editor, map compiler, model viewer, and shader editor and viewer. Since its release, hundreds of mods have been submitted to sites such as FileFront. A 66 MB demo of the game was released on May 10, 2002, featuring the same level shown in an incomplete form at E3 2001, which did not feature in the final version. Two patches were subsequently released: version 1.03 and 1.04.
At E3 2002, LucasArts announced that Jedi Outcast would be released on the GameCube and Xbox. On May 31, 2002, LucasArts and Aspyr announced that a Macintosh version of the game would be released. The Mac version was released on November 5, 2002, and the Xbox and GameCube versions on November 19 in North America and three days later in Europe.
On November 15, 2006, LucasArts announced that Jedi Outcast would feature with Star Wars: Battlefront, Star Wars: Empire at War, Star Wars: Knights of the Old Republic, Star Wars: Republic Commando and a 14-day trial of Star Wars Galaxies in a compilation release entitled Star Wars: The Best of PC. It was released during the 2006 holiday season.
A Nintendo Switch and PlayStation 4 port was announced on September 4, 2019; with a release date of September 24, 2019.
Source-code release
Following Disney's decision to close LucasArts on April 3, 2013, the developers at Raven Software released the source code for the game on SourceForge under GNU GPL-2.0-only. A few days after release, the source code disappeared from SourceForge without explanation. SourceForge later explained to media outlet Kotaku that Raven Software had requested its removal. Kotaku speculated this was due to the presence of licensed code, such as for the Bink Video format from Rad Game Tools, that was not intended to be made public.
However, prior to the removal a fork called "OpenJK" was created on GitHub with the problematic code parts already excised. The goal of the OpenJK community project is to "maintain and improve Jedi Academy + Jedi Outcast released by Raven Software." This includes the unification of the Jedi Outcast and Jedi Academy engines, maintaining the compatibility of previous releases, and providing source ports for Mac and Linux. Daily builds for Windows and early Linux builds are also available.
Reception
PC
In the United States, Jedi Outcasts computer version sold 390,000 copies and earned $17 million by August 2006. At the time, this led Edge to declare it the country's 39th-best-selling computer game, and best-selling Jedi Knight computer title, released since January 2000. Combined sales of all Jedi Knight computer games released during the 2000s, including Jedi Outcast, reached 920,000 units in the United States by August 2006. Jedi Outcasts computer version also received a "Silver" sales award from the Entertainment and Leisure Software Publishers Association (ELSPA), indicating sales of at least 100,000 copies in the United Kingdom; and a "Gold" certification from the Verband der Unterhaltungssoftware Deutschland (VUD), for sales of at least 100,000 units across Germany, Austria and Switzerland. The PC version holds an aggregate score of 89 out of 100 on Metacritic, based on thirty-five reviews.
Game Informer scored the game 9.5 out of 10, with reviewer Andrew Reiner calling it "the quintessential Star Wars game and a feast for all FPS devotees." He also wrote, "without question, Jedi Outcast is the most enjoyable and accomplished Star Wars game yet." He was also extremely impressed with multiplayer, calling it "a highly addictive, over-the-top experience that all FPS players must see to believe." Brian Gee of Game Revolution gave the game an A-. He praised how authentically Star Wars the game felt, as well as the lightsaber interface, writing "I've played a ton of Star Wars games in my time, and I don't think any of them can match the excitement and fun of lightsaber play in Jedi Outcast."
IGN were also impressed, scoring the game 9 out of 10 and giving it an "Editor's Choice" award. Reviewer Steve Butts lauded the "tightly written, mature plot," and wrote "not only is this one of the greatest Star Wars games I've ever played, it's one of the best action games period." He also commended the "fantastic" graphics and "intelligent" level design, although he was critical of the puzzles, the lightsaber interface, and complained that "the game starts too slowly." GameSpot also scored the game 9 out of 10 and, like IGN, also gave it an "Editor's Choice" award. Reviewer Amer Ajami echoed IGN's criticisms of the "slow start" and "too much puzzle-solving." However, he was very impressed with the interface, arguing "never before has melee combat in a shooter been so effectively executed," and he concluded by saying "the game's strong points - especially its combat - overshadow whatever problems Jedi Outcast may have early on," calling it "simply one of the easiest games to recommend this year."
Game Over Online gave the game 93%, calling it "an intriguing juxtaposition of pieces of incredibly intense FPS action that had me on the edge of my seat combined with puzzle-like sections of such opacity that they made me want to kill myself." Entertainment Weekly gave it an A, stating that "the Force is strong with this one." Maxim, however, gave it an eight out of ten and stated that "The Force is strong with this polished first-person shooter."
In a negative review, X-Play criticized the game as a "disturbance in the Force." Although reviewer Jason D'Aprile called the story "pretty good," the graphics "fantastic" and the audio effects "just right," he complained that the level design "succumbs to the Dark Side," citing "illogical and frustrating situations." He also considered the multiplayer mode "not very impressive" and gave the game 2 out of 5 stars. Eurogamer were also underwhelmed, scoring the game 7 out of 10. They criticized the slow start, the lightsaber interface and the AI. However, they did commend the atmosphere and some of the level design. They concluded that "Jedi Outcast is a rather patchy game, with moments of genius let down by a lack of consistency. The guns are derivative and rarely used once you get hold of your lightsaber, the melee combat clumsy and chaotic, the AI poor and some of the locations derivative and poorly designed."
In June 2007, GameTrailers's list of the 10 best Star Wars games saw Jedi Outcast rank at number one, with the editorial commenting that "This was certainly not the first time players had the opportunity to play as a Jedi in a video game, it was the first time you actually felt like one." They also praised the lightsaber and Force combat systems, a story which they felt was consistent with the films, and the cameos from some key Star Wars characters (namely Luke Skywalker and Lando Calrissian).
Jedi Outcast was nominated for Computer Gaming Worlds 2002 "Action Game of the Year" award, which ultimately went to Medal of Honor: Allied Assault. The editors wrote, "Jedi Knight II is a blast, and the force powers and lightsaber control are perfectly executed." However, they found it too lacking in originality and consistency to win the award.
Xbox and GameCube
The Xbox version holds an aggregate score of 81 out of 100 on Metacritic, based on twenty-four reviews. The GameCube version holds scores of 75 out of 100 on Metacritic, based on seventeen reviews.
The Xbox version was described as "truly fantastic" by the Official Xbox Magazine, who scored it 9 out of 10. IGN were also impressed, scoring it 8.8 out of 10 and calling it "the best Star Wars experience on the Xbox". Reviewer Aaron Boudling praised how the controls had been mapped onto the Xbox Controller, but was highly critical of the game's lack of support for Xbox Live, and the absence of an online multiplayer mode. He also felt the graphics in the cutscenes were significantly weaker than the PC version. GameSpot scored it 8.3 out of 10, with reviewer Amer Ajami saying the port "retains the essence of what it must feel like to be an all-powerful Jedi, which was so well conveyed in the original PC game, without sacrificing much in the way of graphics, playability, or overall value. In fact, the game's complicated control scheme is better served on an Xbox controller than it is in the PC's typical mouse and keyboard setup." However, he too criticized the lack of Xbox Live support and the poor graphical quality of the cutscenes. Game Revolution gave the game a B+, praising the controls and atmosphere, but criticizing the AI and lack of Xbox Live support. Eurogamer were as unimpressed with the Xbox version as they had been with the PC version, scoring it 6 out of 10. Reviewer Kristan Reed found a lot of the problems with the game to be inherent to the original, not problems with the port - a slow beginning, poor AI, bad level design, repetitive puzzles. Reed called the lack of Xbox Live support "plain unacceptable" and concluded "Jedi Outcast is certainly an entertaining package when it gets things right, but it's also riddled with design flaws, technical flaws and various minor irritations that conspire to detract from your overall enjoyment. Given that it's also not Live (or even LAN) enabled, Xbox owners don't even have that crumb of comfort to elevate its status."
GameSpot scored the GameCube version 8.2 out of 10. As with the Xbox version, Ajami was critical of the cutscene graphics, but concluded that "Jedi Knight II: Jedi Outcast makes a fine addition to the Nintendo console's gradually growing list of action games." GameSpy gave the same version three-and-a-half stars out of five. Reviewer Scott Steinberg was critical of the game's slow start and the nature of some of the puzzles. Of the port, he wrote "This game was an award-winning game on a prior platform, so you'd be in the right to demand more. But from the perspective of a GameCube owner just getting your first taste of the galactic goodie that is the Jedi Knight series, things could have been a lot worse [...] Jedi Knight II: Outcast stands as a respectable effort, but disappointing reminder that when you swap content from platform to platform, something often gets lost in the transition." IGN scored it 7.2 out of 10, with reviewer Aaron Boudling arguing that the graphics and controls were significantly poorer than in either the PC or Xbox versions. Of the controls, Boulding wrote "While the Xbox version allows you to map your favorite moves to the black and white buttons for quick access, GameCube owners are out of luck and will have to fumble around with the D-Pad in the heat of battle." He concluded "It's obvious that no work at all went into trying to get this game to properly run on GameCube. The framerate is erratic and constantly alternates between smooth and choppy, further complicating the already clunky control system [...] If you're a multi-console owner, get the superior Xbox version to see a better realized port." Nintendo World Report scored it 6 out of 10, also citing the inferior controls and poorer graphics than the PC and Xbox versions; "Jedi Outcast ends up being a pretty terrible port of an excellent PC title. The trick to enjoying the GameCube incarnation will be coming to terms with the control and graphic inadequacies in the game and then being able to make yourself get over the "hump" so that you can enjoy the meat and potatoes of the whole experience."
It was a runner-up for GameSpots annual "Best Shooter on GameCube" award, which went to TimeSplitters 2.
See also
List of formerly proprietary software
List of open source games
References
External links
Official site via Internet Archive
Jedi Outcast source code mirror at GitHub
2002 video games
Activision games
Aspyr games
Commercial video games with freely available source code
First-person shooters
Id Tech games
LucasArts games
Classic Mac OS games
MacOS games
Multiplayer and single-player video games
Multiplayer online games
GameCube games
Nintendo Switch games
PlayStation 4 games
Open-source video games
Raven Software games
Jedi Outcast
Third-person shooters
Video game sequels
Video games developed in the United States
Windows games
Xbox games |
null | null | President pro tempore of the United States Senate | eng_Latn | The president pro tempore of the United States Senate (often shortened to president pro tem) is the second-highest-ranking official of the United States Senate. Article One, Section Three of the United States Constitution provides that the vice president of the United States is the president of the Senate (despite not being a senator), and mandates that the Senate must choose a president pro tempore to act in the vice president's absence. Unlike the vice president, the president pro tempore is an elected member of the Senate, able to speak or vote on any issue. Selected by the Senate at large, usually by a resolution which is adopted by unanimous consent without a formal vote, the president pro tempore has enjoyed many privileges and some limited powers. During the vice president's absence, the president pro tempore is empowered to preside over Senate sessions. Except when necessary or to highlight important votes, the vice president and the president pro tempore rarely preside; instead, the duty of presiding officer is rotated among junior U.S. senators of the majority party to give them experience in parliamentary procedure.
Since 1890, the most senior U.S. senator in the majority party has generally been chosen to be president pro tempore and holds the office continuously until the election of another. This tradition has been observed without interruption since 1949. Since the enactment of the current Presidential Succession Act in 1947, the president pro tempore is third in the line of succession to the presidency, after the vice president and the speaker of the House of Representatives and ahead of the secretary of state.
The current president pro tempore of the Senate is Patrick Leahy of Vermont. He was sworn in on January 20, 2021, during the 117th Congress, when the Democratic Party gained control of the Senate. He previously held the position from 2012 to 2015.
Power and responsibilities
Although the position is in some ways analogous to the speaker of the House of Representatives, the powers of the president pro tempore are far more limited. In the Senate, most power rests with the majority leader and other individual senators, but as the chamber's presiding officer, the president pro tempore is authorized to perform certain duties in the absence of the vice president, including ruling on points of order. Additionally, under the 25th Amendment to the Constitution, the president pro tempore and the speaker are the two authorities to whom declarations must be transmitted that the president is unable to perform the duties of the office, or is able to resume doing so. The president pro tempore is third in the line of presidential succession, following the vice president and the speaker, and consequently is one of the few members of Congress entitled to a full-time security detail. Additional duties include appointment of various congressional officers, certain commissions, advisory boards, and committees and joint supervision of the congressional page school. The president pro tempore is the designated legal recipient of various reports to the Senate, including War Powers Act reports under which they, jointly with the speaker, may require the president to call Congress back into session. The officeholder is an ex officio member of various boards and commissions. With the secretary and sergeant at arms, the president pro tempore maintains order in Senate portions of the Capitol and Senate buildings.
History
Position established
The office of president pro tempore was established by the Constitution of the United States in 1789. Between 1792 and 1886, the president pro tempore was second in the line of presidential succession, following the vice president and preceding the speaker. Through 1891, the president pro tempore was appointed on an intermittent basis only, when the vice president was not present to preside over the Senate, or at the adjournment of a session of Congress.
The first president pro tempore, John Langdon, was elected on April 6, 1789, serving four separate terms between 1789 and 1793. "More than twelve senators held the office during the Senate's first decade," presiding over sessions, signing legislation, and performing routine administrative tasks.
Whenever the office of the vice presidency was vacant, as it was on ten occasions between 1812 and 1889, the office garnered heightened importance, for although he did not assume the vice presidency, the president pro tempore stood next in line for the presidency. Before the ratification of the Twenty-fifth Amendment in 1967, a vacancy in the vice presidency could be filled only by a regular election; several individuals who served during these vacancies were referred to informally as "acting vice president."
On three occasions during the 19th century, the Senate was without both a president and a president pro tempore:
from July 9 to July 11, 1850, following Millard Fillmore's accession to the presidency upon the death of Zachary Taylor, until William R. King was elected president pro tempore;
from September 19 to October 10, 1881, following Chester Arthur's accession to the presidency upon the death of James A. Garfield, until Thomas F. Bayard was elected president pro tempore;
from November 25 to December 7, 1885, following the death of Vice President Thomas A. Hendricks, until John Sherman was elected president pro tempore.
When President Andrew Johnson, who had no vice president, was impeached and tried in 1868, Senate President pro tempore Benjamin Franklin Wade was next in line to the presidency. Wade's radicalism is thought by many historians to be a major reason why the Senate, which did not want to see Wade in the White House, acquitted Johnson.
Vice President Henry Wilson died on November 22, 1875. Senator Thomas W. Ferry, being President pro tempore of the Senate, was next in the line of presidential succession, and remained so until March 4, 1877. As acting president of the senate, he presided over the 1876 impeachment trial of U.S. Secretary of War William Belknap and the meetings of the Electoral Commission created by Congress to resolve the disputed 1876 presidential election. Still president pro tempore at that time, he would have temporally become the acting president had the Electoral College vote not been certified by March 4, 1877; Congress certified Rutherford B. Hayes as the winner of the Electoral College vote on March 2.
The president pro tempore and the speaker of the House were removed from the presidential line of succession in 1886. Both were restored to it in 1947, though this time with the president pro tempore following the speaker.
William P. Frye served as president pro tempore from 1896 to 1911 (the 54th through the 62nd Congress), a tenure longer than anyone else. He resigned from the position due to ill health shortly before his death. Electing his successor proved difficult, as Senate Republicans, then in the majority, were split between progressive and conservative factions, each promoting its own candidate. Likewise, the Democrats proposed their own candidate. As a result of this three-way split, no individual received a majority vote. It took four months for a compromise solution to emerge: Democrat Augustus Bacon served for a single day, August 14, 1911, during the vice president's absence. Thereafter, Bacon and four Republicans—Charles Curtis, Jacob Gallinger, Henry Cabot Lodge, and Frank Brandegee—alternated as president pro tempore for the remainder of the Congress.
Modern era
In January 1945, the 79th Congress elected Kenneth McKellar, who at the time was the senator with the longest continuous service, to be president pro tempore. Since then, it has become customary for the majority party's senior member to hold this position. Arthur Vandenberg (from 1947 to 1949) was the last president pro tempore not to be the senior member of the majority party, aside from the single day in December 1980 accorded Milton Young, who was the retiring senior member of the Republican Party, which would hold the majority in the incoming 97th Congress.
Three presidents pro tempore have gone on to be elected to the office of vice president—John Tyler, William R. King, and Charles Curtis—of whom Tyler is the only one to become president, assuming the office in April 1841 upon the death of William Henry Harrison.
Related officials
Acting president pro tempore
While the president pro tempore does have other official duties, the holders of the office have, like the vice president, over time ceased presiding over the Senate on a daily basis, owing to the mundane and ceremonial nature of the position. Furthermore, as the president pro tempore is usually the most senior senator of the majority party, they most likely also chair a major Senate committee and has other significant demands on their time. Therefore, the president pro tempore has less time now than in the past to preside daily over the Senate. Instead, junior senators from the majority party are designated acting president pro tempore to preside over the Senate. This allows junior senators to learn proper parliamentary procedure. The acting president pro tempore is usually reappointed daily by the president pro tempore.
Permanent acting president pro tempore
In June 1963, because of the illness of president pro tempore Carl Hayden, Lee Metcalf was designated permanent acting president pro tempore. No term was imposed on this designation, so Metcalf retained it until he died in office in 1978.
Deputy president pro tempore
The ceremonial post of deputy president pro tempore was created for Hubert Humphrey, a former vice president, in 1977 following his losing bid to become the Senate majority leader. The Senate resolution creating the position stated that any former president or former vice president serving in the Senate would be entitled to this position. However, since Humphrey, none have served.
George J. Mitchell was elected deputy president pro tempore in 1987, because of the illness of president pro tempore John C. Stennis, similar to Metcalf's earlier designation as permanent acting president pro tempore. The office has remained vacant since 1989 and no senator other than Humphrey and Mitchell has held it since its creation. Mitchell is the only person to have served as deputy president pro tempore who was neither a former president nor former vice president of the United States.
The post is largely honorary and ceremonial, but comes with a salary increase. By statute, the compensation granted to the position holder equals the rate of annual compensation paid to the president pro tempore, majority leader, and minority leader.
President pro tempore emeritus
Since 2001, the honorary title of president pro tempore emeritus has been given to a senator of the minority party who has previously served as president pro tempore. The position has been held by Strom Thurmond (R-South Carolina) (2001–2003), Robert Byrd (D-West Virginia) (2003–2007), Ted Stevens (R-Alaska) (2007–2009), Patrick Leahy (D-Vermont) (2015–2021), and Chuck Grassley (R-Iowa) (2021–present). From 2009 to 2015, no senator met the requirements for the position.
The position was created for Thurmond when the Democratic Party regained a majority in the Senate in June 2001. With the change in party control, Democrat Robert Byrd of West Virginia replaced Thurmond as president pro tempore, reclaiming a position he had previously held from 1989 to 1995 and briefly in January 2001. Thurmond's retirement from the Senate on January 3, 2003, coincided with a change from Democratic to Republican control, making Stevens president pro tempore and Byrd the second president pro tempore emeritus. In 2007, Byrd returned as president pro tempore, and Stevens became the third president pro tempore emeritus, when the Democrats gained control of the Senate. Although a president pro tempore emeritus has no official duties, they are entitled to an increase in staff, and advise party leaders on the functions of the Senate.
The office's accompanying budget increase was removed toward the end of the 113th Congress, shortly before Patrick Leahy was to become the first holder of the office in six years.
Salary
The salary of the president pro tempore for 2012 was $193,400, equal to that of the majority leaders and minority leaders of both houses of Congress. If there is a vacancy in the office of vice president, then the salary would be the same as that of the vice president.
See also
Seniority in the United States Senate
References
External links
United States
United States Senate
Senior legislators |
null | null | Lara Croft: Tomb Raider | eng_Latn | Lara Croft: Tomb Raider is a 2001 action adventure film based on the Tomb Raider video game series featuring the character Lara Croft, portrayed by Angelina Jolie. An international co-production between the United States, the United Kingdom, Japan and Germany, it was directed by Simon West and revolves around Lara Croft trying to obtain ancient artifacts in competition with the Illuminati.
The film was released on June 15, 2001, and received generally negative reviews from critics, although Angelina Jolie was praised for her performance. Tomb Raider was the highest-grossing film on its opening weekend. A sequel, Lara Croft: Tomb Raider – The Cradle of Life, was released in 2003.
Plot
Adventurer Lara Croft defeats a robot in an Egyptian tomb, revealed to be a training exercise arena in her family manor, where she lives with her technical assistant Bryce and butler Hilary. In Venice, as the first phase of a planetary alignment begins, the Illuminati search for a key to rejoin halves of a mysterious artifact, "the Triangle," which must be completed by the final phase, a solar eclipse. Manfred Powell assures the cabal that the artifact is almost ready, but has no real idea of its location.
Lara's father Lord Richard Croft, long missing and presumed dead, appears to her in a dream. Lara awakens to a mysterious ticking, and finds a strange clock hidden inside the manor. On her way to consult a friend of her father's, Wilson, Lara crosses paths with Alex West, an American associate and fellow adventurer. Lara shows Wilson the clock, and he puts her in touch with Powell. Lara shows Powell photographs of the clock, which he claims not to recognize.
That night, armed commandos invade the house and steal the clock, bringing it to Powell. The next morning, a prearranged letter from Lara's father arrives, explaining that the clock is the key to retrieving the halves of the Triangle of Light, an ancient object with the power to control time. After misuse of its power destroyed an entire city, the Triangle was separated: one half was hidden in a tomb at Angkor Cambodia, and the other in the ruined city located at Ukok Plateau, Siberia. Her father tasks her to find and destroy both pieces before the Illuminati can exploit the Triangle's power.
In Cambodia, Lara finds Powell, who has hired West, and his commandos already at the temple. West solves part of the temple's puzzle, and Powell prepares to insert the clock at the moment of alignment. Lara, realizing they are mistaken, finds the correct keyhole; with only seconds left, Lara persuades Powell to throw her the clock. She unlocks the first piece of the Triangle, and the statues of the temple come to life and attack the intruders. West, Powell, and his remaining men flee with the clock, leaving Lara to defeat an enormous six-armed guardian statue. She escapes with the first piece; recovering at a Buddhist monastery, she arranges a meeting with Powell.
In Venice, Powell proposes a partnership to find the Triangle, and informs Lara that her father was a member of the Illuminati, and offers to use the Triangle's power to resurrect him; though reluctant, she agrees to join forces. Lara and Bryce travel with Powell, West, and the leader of the Illuminati to Siberia. Entering the tomb, they discover a giant orrery, which activates as the alignment nears completion. Lara retrieves the second half of the Triangle, and Powell kills the Illuminati's leader to restore the Triangle himself, but the halves will not fuse. Realizing Lara knows the solution, Powell kills West to persuade her to complete the Triangle to restore West's and her father's lives. Lara complies, but seizes the Triangle herself.
In a "crossing" of time, Lara faces the memory of her father, who urges her to destroy the Triangle for good rather than save his life. Returning to the tomb, Lara manipulates time to save West and stab Powell, and destroys the Triangle. The tomb begins to collapse, and all flee but the wounded Powell, who reveals to Lara that he murdered her father. After a hand-to-hand fight, she kills Powell, retrieving her father's pocket watch and escaping the tomb.
Back in her manor, Lara visits her father's memorial and finds that Bryce has reprogrammed the robot, and Hilary presents her with her pistols, which she takes with a smile.
Cast
Angelina Jolie as Lara Croft
Iain Glen as Manfred Powell
Jon Voight as Lord Richard Croft
Daniel Craig as Alex West
Noah Taylor as Bryce
Richard Johnson as Distinguished Gentleman
Chris Barrie as Hillary
Julian Rhind-Tutt as Mr. Pimms
Leslie Phillips as Wilson
Henry Wyndham as Boothby's Auctioneer
Olegar Fedoro as Russian Commander
Production
Development
Tomb Raider went through many drafts and several writers, which resulted in production delays. In 1998, writer Brent V. Friedman, who had co-written Mortal Kombat: Annihilation the year before, penned an unproduced Tomb Raider script. Producer and screenwriter Steven E. de Souza, who wrote and directed the 1994 video game film Street Fighter, penned an early draft of the Tomb Raider script in 1999, but it was rejected by Paramount. The final draft of the script was attributed to five writers, including director Simon West.
Financing
Lara Croft was financed through Tele München Gruppe (TMG), a German tax shelter. The tax law of Germany allowed investors to take an instant tax deduction even on non-German productions and even if the film has not gone into production. By selling them the copyright for $94 million and then buying it back for $83.8 million, Paramount Pictures made $10.2 million. The copyright was then sold again to Lombard Bank, a British investment group and a further $12 million was made. However, to qualify for Section 48 tax relief, the production must include some UK filming and British actors, which was acceptable for a film partially set in the United Kingdom. Presales to distributors in Japan, Britain, France, Germany, Italy and Spain made a further $65 million. Showtime paid $6.8 million for premium cable television rights. In total, $94 million was put together.
Casting
The announcement of the film generated significant discussion about who would be cast to play Lara Croft. Numerous actresses (and non-actresses) were rumoured to be on the shortlist to play her and countless others declared their interest in the role, most notably Jennifer Love Hewitt, Famke Janssen, Jennifer Lopez, Rhona Mitra, Elizabeth Hurley, Ashley Judd, Sandra Bullock, Catherine Zeta-Jones, Diane Lane, Demi Moore and Denise Richards, with many considering the latter the favourite to win the role.
The casting of Jolie was controversial among many fans of the Tomb Raider series, who felt she wasn't physically appropriate enough to play the large-breasted heroine; others complained about an American actress being hired to play a British character; others cited Jolie's tattoos and well-publicised controversial personal life. Director Simon West dismissed these concerns and said, in reference to Jolie's penchant for sexual knife play, "it was always Angelina. I mean, Lara sleeps with knives and doesn't take shit from anybody. That's [Angelina] down to a tee." Jolie wore a padded bra to increase her bust size when playing Lara. As she explained to NY Rock in June 2001: "C'mon, I'm not so flat chested to begin with. When I wear a tight T-shirt, I look a certain way. So it wasn't like we had to completely change me. You know, we just had to enhance me a little. I'm a 36C. Lara, she's a 36D. And in the game, she's a double D, so we took her down some. But we did give her a bit of padding there. For me, it was simply one size. So it was like having a padded bra. But no, I am not flat chested anyway. So we still made it Lara Croft, but we didn't go to any extremes. And Lara doesn't apologize for herself, and for having that, you know, recognizable shape. So I'm not going to apologize for her either."
The film marked the feature film debut of television actor Chris Barrie, known for his role of Arnold Rimmer in the BBC science fiction comedy series Red Dwarf. English actor Daniel Craig adopts an American accent for the role of Alex West whilst Jolie, being American herself, takes on an English accent. Jon Voight, Angelina Jolie's father, plays Richard Croft, Lara's father in the film.
Filming
Principal photography for Lara Croft: Tomb Raider took place from July 30 to November 30, 2000. Portions of the film were shot on location at the Ta Prohm temple, located in Angkor, Siem Reap Province, Cambodia. The film was the first major motion picture to be shot in Cambodia since Lord Jim in 1964, following the country's occupation by the Khmer Rouge regime. In addition to on-location shooting, a majority of the film's production also took place on the 007 Stage at Pinewood Studios. Hatfield House in Hertfordshire was used as Croft's home in the film.
Post-production
Simon West's first director's cut of the film was around 130 minutes long, and that was before it went through re-shoots some time later after principal photography was finished. West was removed from working on the film in post due to some problems between him and Paramount, however he did return to work on re-shoots. Editor Stuart Baird was brought in by Paramount to re-edit the entire film. He was well known for coming in at last minute to help movie studios by fixing up their films which went through troubled productions and had problems with earlier cuts. Baird was promised the job of directing Star Trek: Nemesis by Paramount in exchange for re-editing original 130 minute cut of this film and John Woo's original 210 minute cut of Mission: Impossible 2. Due to all the work on post production taking longer than anyone thought it would, and how rushed it was, some of the major effects were apparently left unfinished by the time the film was released in theaters. Baird ended up re-editing the film down to 88 minutes. One of the bigger cuts on the film was changing the original score. Originally, Nathan McCree was hired to compose the score, since he already composed the music for some of Tomb Raider games, but Paramount insisted on more well known movie composer. Greg Hale Jones started to work on the score with Peter Afterman in November 2000, right after production was finished, and Danny Elfman composed the main theme for the film. Jones later said how once West was fired from the film, his score and Elfman's theme were shelved. Michael Kamen was then hired and he submitted couple demos, although according to Tales From Development Hell by David Hughes, Kamen did in fact compose full score which was rejected once the film was re-edited by Baird. Graeme Revell was then brought in and he had to compose sixty minutes of music inside ten days. Some earlier reports about the film mentioned John Powell as one of original choices for composer, and Fatboy Slim too. Jerry Goldsmith was also attached to score the film at one point, but he couldn't do it due to problems with his health at the time.
Despite large amount of film that was cut out, only four deleted scenes, in total seven minutes long, were included as extras on film's DVD and Blu-Ray releases. Trailers for the film do show at least couple more deleted scenes. At the time of movie's release in July 2001, Simon West said in an interview how he already prepared alternate version of the film for DVD release which would include lot of deleted scenes, but this version was never released. Angelina Jolie's nude scenes were also cut out for PG-13 rating. The shower scene for example originally had actual nude shots of Lara, and next scene where she throws her towel away originally showed some rear nudity from her, however according to Jolie she used body double for her nude shots. One of the special features of the film, Digging Into Tomb Raider, shows filming of another deleted scene where Lara is showering but this one takes place in some outside location. Some of the earlier scripts for the film included more scenes focusing on Lara's sex appeal, like Lara flashing her breasts to some villains as distraction (Steven E. de Souza February 1999 script), using her breasts to crush giant milipede (Mike Werb & Michael Colleary November 1999 script), and many similar scenes (reportedly including her showering, bathing, skinny dipping, losing her bikini while fighting some creature...), but probably due to possible rating problems, this was changed and toned down with each new script.
Music
Soundtrack
Lara Croft: Tomb Raider is a 2001 soundtrack album to the film. The various artists soundtrack was released June 15, 2001. The Score was later released on June 25, 2001. The movie also featured the song "Lila" by Vas and a Piano rendition of "Largo" from Johann Sebastian Bach's Harpsichord Concerto no. 5 performed by Hae-won Chang. These were not featured on the soundtrack. Also used in the movie were elements of "Elevation (Influx Remix)" by U2. This was uncredited.
Score
Graeme Revell composed the soundtrack in less than two weeks, following failed attempts by other composers.
The CD was released through Elektra Entertainment. As noted by Revell, the tracks were mislabeled after failed attempts to stop the pressings. For example, the opening track includes both the Main Titles and Lara Croft at Home cues together. The resulting score caused the composer himself to issue an apology "for the poor listening experience" through his website. The track list was later revised.
Release
Home media
Lara Croft: Tomb Raider was released on DVD and VHS on November 13, 2001; a Blu-ray release followed on June 3, 2008. A 4K UHD Blu-ray release followed on February 27, 2018.
On the North American video rental charts, the film grossed in DVD rental revenue, .
Reception
Box office
Tomb Raider was a box office success. The movie debuted at number one with $48.2 million, ahead of Atlantis: The Lost Empire and Shrek, giving Paramount its second-best debut and the fifth-highest debut of 2001. It beat the opening record for a film featuring a female protagonist ($42.3 million for Scary Movie) as well as the opening record for a video game adaptation ($31 million for Pokémon: The First Movie), and is one of the highest grossing video game to film adaptations. The movie has grossed a total of $274,703,340 worldwide.
Critical response
On review aggregator Rotten Tomatoes, Lara Croft: Tomb Raider earned positive reviews from 20% of 161 critics, with an average rating of 4/10. The site's consensus is "Angelina Jolie is perfect for the role of Lara Croft, but even she can't save the movie from a senseless plot and action sequences with no emotional impact". According to Metacritic, which assigned the film a weighted average score of 33 out of 100 based on reviews from 31 critics, the film received "generally unfavorable reviews". Audiences surveyed by CinemaScore gave the film a grade B on scale of A to F.
IGN gave the movie the lowest score, a 0.0 ("Disaster") rating, condemning everything from character performances to the ending.
Todd McCarthy of Variety said "[the film] has the distinction of being a major motion picture that’s far less imaginative, and quite a bit more stupid, than the interactive game it’s based on." McCarthy praises Jolie but says "everything else about this frenetic production is flat and unexciting."
A positive review came from Roger Ebert, who awarded the film three out of four stars and said, Lara Croft Tomb Raider' elevates goofiness to an art form. Here is a movie so monumentally silly, yet so wondrous to look at, that only a churl could find fault."
Accolades
The film was nominated for two MTV Movie Awards, these awards included: Best Female Performance and Best Fight scene, but the film lost to Moulin Rouge! and Rush Hour 2 respectively. The film was also nominated for Teen Choice Award for Choice Movie - Drama. Angelina Jolie was nominated for the Worst Actress Golden Raspberry Award for her role in the film, but she lost to Mariah Carey in Glitter.
Themes
Director Simon West would comment a decade after its release that the creation of Lara Croft was influenced by a film market that "wasn't used to women leading summer blockbusters". This factor influenced his decision to cast Angelina Jolie who was not well known at the time, and not the studio's first choice (in contrast to Catherine Zeta-Jones, Ashley Judd, and Jennifer Lopez). West said that his decision to cast Jolie lay in the fact that "there hadn't been a female lead of an action-adventure film that had carried a film [by herself recently], and Angelina wasn't as big as some of the other actresses that were up for the part, who'd done bigger films and had a longer track record and bigger box-office grosses... Some of their [images] were safer than Angelina's, whose was quite dangerous. She had all sorts of thing written about her—some obviously not true. She was a young woman experimenting." While Lara Crofts box office totals were the highest for a female-led action film at the time, and the film inspired theme park rides and led to a sequel, West stated in 2018 that "at the time, the studio was incredibly nervous at what the outcome could have been. I'm surprised it's taken so long [for other female-fronted action stories to rise up], because I thought that two or three years after, there'd be 10 other movies like it cashing in on its success ...[b]ut it's amazing how things work so slowly. But finally The Hunger Games and Wonder Woman have caught up!"
Other media
Sequel
Angelina Jolie returned in the sequel Lara Croft: Tomb Raider – The Cradle of Life. While it was viewed as a critical improvement over its predecessor, it did not repeat its financial success, grossing $156 million compared to the previous installment's $274 million.
Reboot
GK Films first acquired the rights to reboot the film in 2011. In April 2016, MGM and GK Films announced a reboot of the film starring Alicia Vikander as Lara Croft with Roar Uthaug directing. It was released March 16, 2018.
See also
List of films featuring eclipses
Notes
References
External links
2001 films
2000s action adventure films
2000s fantasy adventure films
American action adventure films
American fantasy adventure films
American films
German films
Japanese films
British films
British action adventure films
British fantasy adventure films
English-language films
English-language German films
English-language Japanese films
Live-action films based on video games
Films directed by Simon West
Films produced by Lawrence Gordon
Films scored by Graeme Revell
Films set in Cambodia
Films set in Egypt
Films set in Russia
Films set in Venice
Films shot in Hertfordshire
Films shot at Pinewood Studios
Films shot in Iceland
Girls with guns films
Paramount Pictures films
Mutual Film Company films
Tomb Raider films
Treasure hunt films
Films about the Illuminati
Works based on Square Enix video games
Adventure film soundtracks
2001 soundtrack albums
Films with screenplays by Michael Colleary |
null | null | List of Star Trek: The Next Generation episodes | eng_Latn | Star Trek: The Next Generation is an American science fiction television series which aired in syndication from September 1987 through May 1994. It is the second live-action series of the Star Trek franchise and comprises a total of 178 episodes over 7 seasons. The series picks up about 95 years after the original series is said to have taken place. The television episodes are listed here in chronological order by original air date, which match the episode order in each season's DVD set.
The main cast consisted of Patrick Stewart as captain Jean-Luc Picard, Jonathan Frakes as his second-in-command William Riker, Brent Spiner as chief of operations Data, LeVar Burton as chief engineer Geordi La Forge, Marina Sirtis as counselor Deanna Troi, Michael Dorn as chief of security Worf, and Gates McFadden as Dr. Beverly Crusher. McFadden left the show after the first season and was replaced for the second season with Diana Muldaur as Dr. Katherine Pulaski, but returned for the third season and remained with the cast thereafter. Wil Wheaton starred as Wesley Crusher in seasons 1–4, returning for guest appearances in seasons 5 and 7. Denise Crosby played chief security officer Tasha Yar in the first season but her character was killed in the episode "Skin of Evil", returning for a guest appearance in season 3.
The Next Generation cast also appears in four feature films: Generations, First Contact, Insurrection, and Nemesis.
Series overview
Episodes
Both the series' premiere ("Encounter at Farpoint") and finale ("All Good Things...") originally aired as single two-hour presentations, but have since aired as two one-hour episodes.
Season 1 (1987–1988)
Season 2 (1988–1989)
Season 3 (1989–1990)
Season 4 (1990–1991)
Season 5 (1991–1992)
Season 6 (1992–1993)
Season 7 (1993–1994)
Home media
The series was released on VHS cassette tapes, usually with two episodes or a two-part episode on one tape. For some releases the credits of both shows on the tape were combined at the end rather than shown separately.
DVD and Blu-ray
Certain episodes of the series have also been released on eight "best of" collections.
The four Next Generation movies were released together on a Blu-ray set in 2009. They have subsequently been re-released individually.
Three episodes were released on region-free Blu-ray Disc in January 2012. The complete series was digitally remastered in 1080p high definition to be available for Blu-ray releases, syndication on television, and distribution digitally.
Five "Fan Collective" DVD collections were released, each including episodes with a particular theme – Borg, Klingon, Q, Time Travel, and Captain's Log – from the first five live-action Star Trek series, including The Next Generation.
Streaming
The digital distribution of the remastered series has been through services such as iTunes and Amazon Video. Seasons 1 and 2 were released on the US, Canadian and UK iTunes stores in early 2013, in both high and standard definition. Seasons 3 and 4 were released on the US and Canadian iTunes stores in the second half of 2013, and on the UK store in March 2014 - again in both high and standard definition. In early 2014, the first four seasons were added to the Australian iTunes Store in both high and standard definition. Both high and standard definition versions of Season 5 were released digitally in the US via Amazon Video and iTunes in early May 2014, and later in the UK, Canadian and Australian iTunes stores. In December 2014, Season 6 was added in both high and standard definition to the US iTunes Store.
The complete series was added to the Netflix streaming service in the US in 2011. In 2015, the series was updated to the high-definition remastered version of the show. Although the series was available to subscribers on Netflix, Hulu, and Amazon Prime for a number of years, after Paramount launched its own streaming service, it stopped renewing streaming rights to the Star Trek series, and by early 2022, The Next Generation was available for streaming only from Paramount+.
See also
Lists of Star Trek episodes
Notes
References
Bibliography
External links
Memory Alpha – a Star Trek wiki containing detailed episode guides
Episode guide at startrek.com
Next Generation, The
Lists of American science fiction television series episodes |
null | null | Wilhelm II, German Emperor | eng_Latn | Wilhelm II (Friedrich Wilhelm Viktor Albert; 27 January 18594 June 1941), anglicised as William II, was the last German Emperor () and King of Prussia, reigning from 15 June 1888 until his abdication on 9 November 1918. Despite strengthening the German Empire's position as a great power by building a powerful navy, his tactless public statements and erratic foreign policy greatly antagonized the international community and are considered by many to be one of the underlying causes for World War I. When the German war effort collapsed after a series of crushing defeats on the Western Front in 1918, he was forced to abdicate, thereby bringing an end to the House of Hohenzollern's 400-year reign.
Wilhelm II was the son of Prince Frederick William of Prussia and Victoria, Princess Royal. His father was the son of Wilhelm I, German Emperor, and his mother was the eldest daughter of Queen Victoria of the United Kingdom and Prince Albert of Saxe-Coburg and Gotha. Wilhelm's grandfather, Wilhelm I, died in March 1888. His father became Emperor Frederick III, but died just 99 days later, and, in what is still called the Year of the Three Emperors, Wilhelm II ascended the throne of the Second Reich in June 1888.
In March 1890, Wilhelm II dismissed the German Empire's powerful longtime Chancellor, Otto von Bismarck, and assumed direct control over his nation's policies, embarking on a bellicose "New Course" to cement its status as a leading world power. Over the course of his reign, the German colonial empire acquired new territories in China and the Pacific (such as Kiautschou Bay, the Northern Mariana Islands, and the Caroline Islands) and became Europe's largest manufacturer. However, Wilhelm often undermined such progress by making tactless statements and threats towards other countries without first consulting his ministers. Britain became Germany's main enemy when the Kaiser launched a massive expansion of the Imperial German Navy. By 1910 Germany had two allies left: the weak Austria-Hungary and the declining Ottoman Empire.
Wilhelm's reign culminated in Germany's guarantee of military support to Austria-Hungary during the crisis of July 1914, one of the immediate causes for World War I. By this time Wilhelm had lost virtually all decision-making power. Indeed all civilian officials were losing power to the Army's General Staff. By August 1916, a de facto military dictatorship set national policy for the rest of the conflict. Despite emerging victorious over Russia and obtaining significant territorial gains in Eastern Europe, Germany was forced to relinquish all its conquests after a decisive defeat on the Western Front in the fall of 1918. Losing the support of his country's military and many of his subjects, Wilhelm was forced to abdicate during the German Revolution of 1918–1919. The revolution converted Germany from a monarchy into an unstable democratic state known as the Weimar Republic. Wilhelm fled to exile in the Netherlands where he remained during its occupation by Nazi Germany in 1940. He died there in 1941.
Biography
Wilhelm was born in Berlin on 27 January 1859—at the Crown Prince's Palace—to Victoria, Princess Royal "Vicky", the eldest daughter of Britain's Queen Victoria, and Prince Frederick William of Prussia ("Fritz" – the future Frederick III). At the time of his birth, his granduncle, Frederick William IV, was king of Prussia. Frederick William IV had been left permanently incapacitated by a series of strokes, and his younger brother Wilhelm was acting as regent. Wilhelm was the first grandchild of his maternal grandparents (Queen Victoria and Prince Albert), but more importantly, he was the first son of the crown prince of Prussia. Upon the death of Frederick William IV in January 1861, Wilhelm's paternal grandfather (the elder Wilhelm) became king, and the two-year-old Wilhelm became second in the line of succession to Prussia. After 1871, Wilhelm also became second in the line to the newly created German Empire, which, according to the constitution of the German Empire, was ruled by the Prussian king. At the time of his birth, he was also sixth in the line of succession to the British throne, after his maternal uncles and his mother.
Traumatic birth
Shortly before midnight on 26 January 1859, Wilhelm's mother Vicky experienced labour pains, followed by her water breaking, after which Dr. August Wegner, the family's personal physician, was summoned. Upon examining Vicky, Wegner realised the infant was in the breech position; gynaecologist Eduard Arnold Martin was then sent for, arriving at the palace at 10 am on 27 January. After administering ipecac and prescribing a mild dose of chloroform, which was administered by Queen Victoria's personal physician Sir James Clark, Martin advised Fritz the unborn child's life was endangered. As mild anaesthesia did not alleviate Vicky's extreme labour pains, resulting in her "horrible screams and wails," Clark finally administered full anaesthesia. Observing Vicky's contractions to be insufficiently strong, Martin administered a dose of ergot extract, and at 2:45 pm saw the infant's buttocks emerging from the birth canal, but noticed the pulse in the umbilical cord was weak and intermittent. Despite this dangerous sign, Martin ordered a further heavy dose of chloroform so he could better manipulate the infant. Observing the infant's legs to be raised upwards and his left arm likewise raised upwards and behind his head, Martin "carefully eased out the Prince's legs." Due to the "narrowness of the birth canal," he then forcibly pulled the left arm downwards, tearing the brachial plexus, then continued to grasp the left arm to rotate the infant's trunk and free the right arm, likely exacerbating the injury. After completing the delivery, and despite realising the newborn prince was hypoxic, Martin turned his attention to the unconscious Princess Victoria. Noticing after some minutes that the newborn remained silent, Martin and the midwife Fräulein Stahl worked frantically to revive the prince; finally, despite the disapproval of those present, Stahl spanked the newborn vigorously until "a weak cry escaped his pale lips."
Modern medical assessments have concluded Wilhelm's hypoxic state at birth, due to the breech delivery and the heavy dosage of chloroform, left him with minimal to mild brain damage, which manifested itself in his subsequent hyperactive and erratic behaviour, limited attention span and impaired social abilities. The brachial plexus injury resulted in Erb's palsy, which left Wilhelm with a withered left arm about shorter than his right. He tried with some success to conceal this; many photographs show him holding a pair of white gloves in his left hand to make the arm seem longer. In others, he holds his left hand with his right, has his crippled arm on the hilt of a sword, or holds a cane to give the illusion of a useful limb posed at a dignified angle. Historians have suggested that this disability affected his emotional development.
Early years
In 1863, Wilhelm was taken to England to be present at the wedding of his Uncle Bertie (later King Edward VII), and Princess Alexandra of Denmark. Wilhelm attended the ceremony in a Highland costume, complete with a small toy dirk. During the ceremony, the four-year-old became restless. His eighteen-year-old uncle Prince Alfred, charged with keeping an eye on him, told him to be quiet, but Wilhelm drew his dirk and threatened Alfred. When Alfred attempted to subdue him by force, Wilhelm bit him on the leg. His grandmother, Queen Victoria, missed seeing the fracas; to her Wilhelm remained "a clever, dear, good little child, the great favourite of my beloved Vicky".
His mother, Vicky, was obsessed with his damaged arm, blaming herself for the child's handicap and insisted that he become a good rider. The thought that he, as heir to the throne, should not be able to ride was intolerable to her. Riding lessons began when Wilhelm was eight and were a matter of endurance for Wilhelm. Over and over, the weeping prince was set on his horse and compelled to go through the paces. He fell off time after time but despite his tears, was set on its back again. After weeks of this he was finally able to maintain his balance.
Wilhelm, from six years of age, was tutored and heavily influenced by the 39-year-old teacher Georg Ernst Hinzpeter. "Hinzpeter", he later wrote, "was really a good fellow. Whether he was the right tutor for me, I dare not decide. The torments inflicted on me, in this pony riding, must be attributed to my mother."
As a teenager he was educated at Kassel at the Friedrichsgymnasium. In January 1877, Wilhelm finished high school and on his eighteenth birthday received as a present from his grandmother, Queen Victoria, the Order of the Garter. After Kassel he spent four terms at the University of Bonn, studying law and politics. He became a member of the exclusive Corps Borussia Bonn. Wilhelm possessed a quick intelligence, but this was often overshadowed by a cantankerous temper.
As a scion of the royal house of Hohenzollern, Wilhelm was exposed from an early age to the military society of the Prussian aristocracy. This had a major impact on him and, in maturity, Wilhelm was seldom seen out of uniform. The hyper-masculine military culture of Prussia in this period did much to frame his political ideals and personal relationships.
Crown Prince Frederick William was viewed by his son with a deeply felt love and respect. His father's status as a hero of the wars of unification was largely responsible for the young Wilhelm's attitude, as were the circumstances in which he was raised; close emotional contact between father and son was not encouraged. Later, as he came into contact with the Crown Prince's political opponents, Wilhelm came to adopt more ambivalent feelings toward his father, perceiving the influence of Wilhelm's mother over a figure who should have been possessed of masculine independence and strength. Wilhelm also idolised his grandfather, Wilhelm I, and he was instrumental in later attempts to foster a cult of the first German Emperor as "Wilhelm the Great". However, he had a distant relationship with his mother.
Wilhelm resisted attempts by his parents, especially his mother, to educate him in a spirit of British liberalism. Instead, he agreed with his tutors' support of autocratic rule, and gradually became thoroughly 'Prussianized' under their influence. He thus became alienated from his parents, suspecting them of putting Britain's interests first. The German Emperor, Wilhelm I, watched as his grandson, guided principally by the Crown Princess Victoria, grew to manhood. When Wilhelm was nearing twenty-one the Emperor decided it was time his grandson should begin the military phase of his preparation for the throne. He was assigned as a lieutenant to the First Regiment of Foot Guards, stationed at Potsdam. "In the Guards," Wilhelm said, "I really found my family, my friends, my interests—everything of which I had up to that time had to do without." As a boy and a student, his manner had been polite and agreeable; as an officer, he began to strut and speak brusquely in the tone he deemed appropriate for a Prussian officer.
In many ways, Wilhelm was a victim of his inheritance and of Otto von Bismarck's machinations. When Wilhelm was in his early twenties, Bismarck tried to separate him from his parents (who opposed Bismarck and his policies) with some success. Bismarck planned to use the young prince as a weapon against his parents in order to retain his own political dominance. Wilhelm thus developed a dysfunctional relationship with his parents, but especially with his English mother. In an outburst in April 1889, Wilhelm angrily implied that "an English doctor killed my father, and an English doctor crippled my arm—which is the fault of my mother", who allowed no German physicians to attend to herself or her immediate family.
As a young man, Wilhelm fell in love with one of his maternal first cousins, Princess Elisabeth of Hesse-Darmstadt. She turned him down, and in time, married into the Russian imperial family. In 1880 Wilhelm became engaged to Augusta Victoria of Schleswig-Holstein, known as "Dona". The couple married on 27 February 1881, and remained married for forty years, until her death in 1921. In a period of ten years, between 1882 and 1892, Augusta Victoria bore Wilhelm seven children, six sons and a daughter.
Beginning in 1884, Bismarck began advocating that Kaiser Wilhelm send his grandson on diplomatic missions, a privilege denied to the Crown Prince. That year, Prince Wilhelm was sent to the court of Tsar Alexander III of Russia in St. Petersburg to attend the coming of age ceremony of the sixteen-year-old Tsarevich Nicholas. Wilhelm's behaviour did little to ingratiate himself to the tsar. Two years later, Kaiser Wilhelm I took Prince Wilhelm on a trip to meet with Emperor Franz Joseph I of Austria-Hungary. In 1886, also, thanks to Herbert von Bismarck, the son of the Chancellor, Prince Wilhelm began to be trained twice a week at the Foreign Ministry. One privilege was denied to Prince Wilhelm: to represent Germany at his maternal grandmother, Queen Victoria's, Golden Jubilee celebrations in London in 1887.
Accession
Kaiser Wilhelm I died in Berlin on 9 March 1888, and Prince Wilhelm's father ascended the throne as Frederick III. He was already suffering from an incurable throat cancer and spent all 99 days of his reign fighting the disease before dying. On 15 June of that same year, his 29-year-old son succeeded him as German Emperor and King of Prussia.
Although in his youth he had been a great admirer of Otto von Bismarck, Wilhelm's characteristic impatience soon brought him into conflict with the "Iron Chancellor", the dominant figure in the foundation of his empire. The new Emperor opposed Bismarck's careful foreign policy, preferring vigorous and rapid expansion to protect Germany's "place in the sun". Furthermore, the young Emperor had come to the throne determined to rule as well as reign, unlike his grandfather. While the letter of the imperial constitution vested executive power in the emperor, Wilhelm I had been content to leave day-to-day administration to Bismarck. Early conflicts between Wilhelm II and his chancellor soon poisoned the relationship between the two men. Bismarck believed that Wilhelm was a lightweight who could be dominated, and he showed scant respect for Wilhelm's policies in the late 1880s. The final split between monarch and statesman occurred soon after an attempt by Bismarck to implement a far-reaching anti-Socialist law in early 1890.
Rift with Bismarck
The young Kaiser allegedly rejected Bismarck's "peaceful foreign policy" and instead plotted with senior generals to work "in favour of a war of aggression". Bismarck told an aide, "That young man wants war with Russia, and would like to draw his sword straight away if he could. I shall not be a party to it."
Bismarck, after gaining an absolute majority in the Reichstag in favour of his policies, decided to push through legislation making his Anti-Socialist Laws permanent. His Kartell, the majority of the amalgamated German Conservative Party and the National Liberal Party, favoured making the laws permanent, with one exception: the police power to expel Socialist agitators from their homes. The Kartell split over this issue and nothing was passed.
As the debate continued, Wilhelm became more and more interested in social problems, especially the treatment of mine workers who went on strike in 1889. He routinely argued with Bismarck in the Council to make it clear where he stood on social policy. Bismarck, in turn, sharply disagreed with Wilhelm's pro-labor union policies and worked to circumvent them. Bismarck, feeling pressured and unappreciated by the young Emperor and undermined by his ambitious advisors, refused to sign a proclamation regarding the protection of workers along with Wilhelm, as was required by the German Constitution.
While Bismarck had previously sponsored landmark social security legislation, by 1889–90, he had become violently opposed to the rise of organized labor. In particular, he was opposed to wage increases, improving working conditions, and regulating labour relations. Moreover, the Kartell, the shifting coalition government that Bismarck had been able to maintain since 1867, had finally lost its majority of seats in the Reichstag.
The final break between the Iron Chancellor and the Monarchy came when Bismarck searched for a new parliamentary majority after his Kartell was voted from power due to the Anti-Socialist Laws fiasco. The remaining powers in the Reichstag were the Catholic Centre Party and the Conservative Party.
In most parliamentary systems, the head of government depends upon the confidence of the parliamentary majority and has the right to form coalitions to maintain a majority of supporters. In a constitutional monarchy, however, the Chancellor also cannot afford to make an enemy of the monarch, who has plenty of means at his or her disposal of quietly blocking a Chancellor's policy objectives. For these reasons, the Kaiser believed that he had the right to be informed before The Iron Chancellor began coalition talks with the Opposition.
In a deeply ironic moment, a mere decade after demonizing Germany's Catholics as traitors during the Kulturkampf, Bismarck decided to start coalition talks with the all-Catholic Centre Party, and invited that party's leader in the Reichstag, Baron Ludwig von Windthorst, to meet with him to begin the negotiations. Despite having a warm relationship with Baron von Windthorst, Kaiser Wilhelm was furious to hear about the plans for coalition talks only after they had already begun.
After a heated argument at Bismarck's estate over the latter's alleged disrespect for the monarchy, Wilhelm stormed out. Bismarck, forced for the first time in his career into a crisis that he could not twist to his own advantage, wrote a blistering letter of resignation, decrying Wilhelm's involvement in both foreign and domestic policy, which was published only after Bismarck's death.
At the opening of the Reichstag on 6 May 1890, the Kaiser stated that the most pressing issue was the further enlargement of the bill concerning the protection of the labourer. In 1891, the Reichstag passed the Workers Protection Acts, which improved working conditions, protected women and children and regulated labour relations.
Wilhelm in control
Dismissal of Bismarck
Bismarck resigned at Wilhelm II's insistence in 1890, at the age of 75. He was succeeded as Chancellor of Germany and Minister-President of Prussia by Leo von Caprivi, who in turn was replaced by Chlodwig, Prince of Hohenlohe-Schillingsfürst, in 1894. Following the dismissal of Hohenlohe in 1900, Wilhelm appointed the man whom he regarded as "his own Bismarck", Bernhard von Bülow.
In foreign policy Bismarck had achieved a fragile balance of interests between Germany, France and Russia — peace was at hand and Bismarck tried to keep it that way despite growing popular sentiment against Britain (regarding colonies) and especially against Russia. With Bismarck's dismissal, the Russians now expected a reversal of policy in Berlin, so they quickly came to terms with France, beginning a process that by 1914 largely isolated Germany.
In later years, Bismarck created the "Bismarck myth"; the view (which some historians have argued was confirmed by subsequent events) that Wilhelm II's successful demand for the Iron Chancellor's resignation destroyed any chance Imperial Germany ever had of stable government and international peace. According to this view, what Kaiser Wilhelm termed "The New Course" is characterised as Germany's ship of state going dangerously off course, leading directly to the carnage of the First and Second World Wars.
In contrast, historian Modris Eksteins has argued that Bismarck's dismissal was actually long overdue. According to Eksteins, the Iron Chancellor, in his need for a scapegoat, had demonized Classical Liberals in the 1860s, Roman Catholics in the 1870s, and Socialists in the 1880s with the highly successful and often repeated refrain, "The Reich is in danger." Therefore, in order to divide and rule, Bismarck ultimately left the German people even more divided in 1890 than they had ever been before 1871.
In appointing Caprivi and then Hohenlohe, Wilhelm was embarking upon what is known to history as "the New Course", in which he hoped to exert decisive influence in the government of the empire. There is debate amongst historians as to the precise degree to which Wilhelm succeeded in implementing "personal rule" in this era, but what is clear is the very different dynamic which existed between the Crown and its chief political servant (the Chancellor) in the "Wilhelmine Era". These chancellors were senior civil servants and not seasoned politician-statesmen like Bismarck. Wilhelm wanted to preclude the emergence of another Iron Chancellor, whom he ultimately detested as being "a boorish old killjoy" who had not permitted any minister to see the Emperor except in his presence, keeping a stranglehold on effective political power. Upon his enforced retirement and until his dying day, Bismarck became a bitter critic of Wilhelm's policies, but without the support of the supreme arbiter of all political appointments (the Emperor) there was little chance of Bismarck exerting a decisive influence on policy.
In the early twentieth century, Wilhelm began to concentrate upon his real agenda: the creation of a German Navy that would rival that of Britain and enable Germany to declare itself a world power. He ordered his military leaders to read Admiral Alfred Thayer Mahan's book, The Influence of Sea Power upon History, and spent hours drawing sketches of the ships that he wanted built. Bülow and Bethmann Hollweg, his loyal chancellors, looked after domestic affairs, while Wilhelm began to spread alarm in the chancellories of Europe with his increasingly eccentric views on foreign affairs.
Promoter of arts and sciences
Wilhelm enthusiastically promoted the arts and sciences, as well as public education and social welfare. He sponsored the Kaiser Wilhelm Society for the promotion of scientific research; it was funded by wealthy private donors and by the state and comprised a number of research institutes in both pure and applied sciences. The Prussian Academy of Sciences was unable to avoid the Kaiser's pressure and lost some of its autonomy when it was forced to incorporate new programs in engineering, and award new fellowships in engineering sciences as a result of a gift from the Kaiser in 1900.
Wilhelm supported the modernisers as they tried to reform the Prussian system of secondary education, which was rigidly traditional, elitist, politically authoritarian, and unchanged by the progress in the natural sciences. As hereditary Protector of the Order of Saint John, he offered encouragement to the Christian order's attempts to place German medicine at the forefront of modern medical practice through its system of hospitals, nursing sisterhood and nursing schools, and nursing homes throughout the German Empire. Wilhelm continued as Protector of the Order even after 1918, as the position was in essence attached to the head of the House of Hohenzollern.
Personality
Historians have frequently stressed the role of Wilhelm's personality in shaping his reign. Thus, Thomas Nipperdey concludes he was:
Historian David Fromkin states that Wilhelm had a love–hate relationship with Britain. According to Fromkin
"From the outset, the half-German side of him was at war with the half-English side. He was wildly jealous of the British, wanting to be British, wanting to be better at being British than the British were, while at the same time hating them and resenting them because he never could be fully accepted by them".
Langer et al. (1968) emphasise the negative international consequences of Wilhelm's erratic personality:
"He believed in force, and the 'survival of the fittest' in domestic as well as foreign politics ... William was not lacking in intelligence, but he did lack stability, disguising his deep insecurities by swagger and tough talk. He frequently fell into depressions and hysterics ... William's personal instability was reflected in vacillations of policy. His actions, at home as well as abroad, lacked guidance, and therefore often bewildered or infuriated public opinion. He was not so much concerned with gaining specific objectives, as had been the case with Bismarck, as with asserting his will. This trait in the ruler of the leading Continental power was one of the main causes of the uneasiness prevailing in Europe at the turn-of-the-century".
Relationships with foreign relatives
As a grandchild of Queen Victoria, Wilhelm was a first cousin of King George V of the United Kingdom, as well as of queens Marie of Romania, Maud of Norway, Victoria Eugenie of Spain, and the empress Alexandra of Russia. In 1889, Wilhelm's younger sister, Sophia, married the future King Constantine I of Greece. Wilhelm was infuriated by his sister's conversion from Lutheranism to Greek Orthodoxy; upon her marriage, he attempted to ban her from entering Germany.
Wilhelm's most contentious relationships were with his British relations. He craved the acceptance of his grandmother, Queen Victoria, and of the rest of her family. Despite the fact that his grandmother treated him with courtesy and tact, his other relatives largely denied him acceptance. He had an especially bad relationship with his Uncle Bertie, the Prince of Wales (later King Edward VII). Between 1888 and 1901 Wilhelm resented his uncle, who despite being an heir-apparent to the British throne, treated Wilhelm not as a reigning monarch, but merely as another nephew. In turn, Wilhelm often snubbed his uncle, whom he referred to as "the old peacock" and lorded his position as emperor over him. Beginning in the 1890s, Wilhelm made visits to England for Cowes Week on the Isle of Wight and often competed against his uncle in the yacht races. Edward's wife, the Danish-born Alexandra, also disliked Wilhelm. Even though Wilhelm had not been on the throne at the time, Alexandra felt anger over the Prussian seizure of Schleswig-Holstein from Denmark in the 1860s, and was also annoyed over Wilhelm's treatment of his mother. Despite his poor relations with his English relatives, when he received news that Queen Victoria was dying at Osborne House in January 1901, Wilhelm travelled to England and was at her bedside when she died, and he remained for the funeral. He also was present at the funeral of King Edward VII in 1910.
In 1913, Wilhelm hosted a lavish wedding in Berlin for his only daughter, Victoria Louise. Among the guests at the wedding were his cousins Tsar Nicholas II of Russia and King George V, and George's wife, Queen Mary.
Foreign affairs
German foreign policy under Wilhelm II was faced with a number of significant problems. Perhaps the most apparent was that Wilhelm was an impatient man, subjective in his reactions and affected strongly by sentiment and impulse. He was personally ill-equipped to steer German foreign policy along a rational course. It is now widely recognised that the various spectacular acts which Wilhelm undertook in the international sphere were often partially encouraged by the German foreign policy elite. There were a number of examples, such as the Kruger telegram of 1896 in which Wilhelm congratulated President Paul Kruger for preventing the Transvaal Republic from being annexed by the British Empire during the Jameson Raid.
British public opinion had been quite favourable towards the Kaiser in his first twelve years on the throne, but it turned sour in the late 1890s. During the First World War, he became the central target of British anti-German propaganda and the personification of a hated enemy.
Wilhelm invented and spread fears of a yellow peril trying to interest other European rulers in the perils they faced by invading China; few other leaders paid attention. Wilhelm used the Japanese victory in the Russo-Japanese War to try to incite fear in the west of the yellow peril that they faced by a resurgent Japan, which Wilhelm claimed would ally with China to overrun the west. Under Wilhelm, Germany invested in strengthening its colonies in Africa and the Pacific, but few became profitable and all were lost during the First World War. In South West Africa (now Namibia), a native revolt against German rule led to the Herero and Namaqua genocide, although Wilhelm eventually ordered it to be stopped.
One of the few times when Wilhelm succeeded in personal diplomacy was when in 1900 he supported the marriage of Archduke Franz Ferdinand of Austria to Countess Sophie Chotek, against the wishes of Emperor Franz Joseph I of Austria.
A domestic triumph for Wilhelm was when his daughter Victoria Louise married the Duke of Brunswick in 1913; this helped heal the rift between the House of Hanover and the House of Hohenzollern that had followed the annexation of Hanover by Prussia in 1866.
Political visits to the Ottoman Empire
In his first visit to Istanbul in 1889, Wilhelm secured the sale of German-made rifles to the Ottoman Army. Later on, he had his second political visit to the Ottoman Empire as a guest of Sultan Abdülhamid II. The Kaiser started his journey to the Ottoman Eyalets with Istanbul on 16 October 1898; then he went by yacht to Haifa on 25 October. After visiting Jerusalem and Bethlehem, the Kaiser went back to Jaffa to embark to Beirut, where he took the train passing Aley and Zahlé to reach Damascus on 7 November. While visiting the Mausoleum of Saladin the following day, the Kaiser made a speech:
On 10 November, Wilhelm went to visit Baalbek before heading to Beirut to board his ship back home on 12 November. In his second visit, Wilhelm secured a promise for German companies to construct the Berlin–Baghdad railway, and had the German Fountain constructed in Istanbul to commemorate his journey.
His third visit was on 15 October 1917, as the guest of Sultan Mehmed V.
Hun speech of 1900
The Boxer Rebellion, an anti-foreign uprising in China, was put down in 1900 by an international force of British, French, Russian, Austrian, Italian, American, Japanese, and German troops. The Kaiser's farewell address to his departing soldiers commanded them, in the spirit of the Huns, to be merciless in battle. Wilhelm's fiery rhetoric clearly expressed his vision for Germany as one of the great powers. There were two versions of the speech. The German Foreign Office issued an edited version, making sure to omit one particularly incendiary paragraph that they regarded as diplomatically embarrassing. The edited version was this:
The official version omitted the following passage from which the speech derives its name:
The term "Hun" later became the favoured epithet of Allied anti-German war propaganda during the First World War.
Eulenberg Scandal
In the years 1906–1909, Socialist journalist Maximilian Harden published accusations of homosexual activity involving ministers, courtiers, army officers, and Wilhelm's closest friend and advisor, Prince Philipp zu Eulenberg. According to Robert K. Massie:Homosexuality was officially repressed in Germany.... It was a criminal offense, punishable by prison, although the law was rarely invoked or enforced. Still, the very accusation could stir moral outrage and bring social ruin. This was especially true at the highest levels of Society. The result was years of highly publicized scandals, trials, resignations, and suicides. Harden, like some in the upper echelons of the military and Foreign Office, resented Eulenberg's approval of the Anglo-French Entente, and also his encouragement of Wilhelm to rule personally. The scandal led to Wilhelm suffering a nervous breakdown, and the removal of Eulenberg and others of his circle from the court. The view that Wilhelm was a deeply repressed homosexual is increasingly supported by scholars: certainly, he never came to terms with his feelings for Eulenberg. Historians have linked the Eulenberg scandal to a fundamental shift in German policy that heightened its military aggressiveness and ultimately contributed to World War I.
Moroccan Crisis
One of Wilhelm's diplomatic blunders sparked the Moroccan Crisis of 1905. He made a spectacular visit to Tangier, in Morocco on 31 March 1905. He conferred with representatives of Sultan Abdelaziz of Morocco. The Kaiser proceeded to tour the city on the back of a white horse. The Kaiser declared he had come to support the sovereignty of the Sultan—a statement which amounted to a provocative challenge to French influence in Morocco. The Sultan subsequently rejected a set of French-proposed governmental reforms and invited major world powers to a conference that advised him on necessary reforms.
The Kaiser's presence was seen as an assertion of German interests in Morocco, in opposition to those of France. In his speech, he even made remarks in favour of Moroccan independence, and this led to friction with France, which was expanding its colonial interests in Morocco, and to the Algeciras Conference, which served largely to further isolate Germany in Europe.
Daily Telegraph Affair
Wilhelm's most damaging personal blunder cost him much of his prestige and power and had a far greater impact in Germany than overseas. The Daily Telegraph Affair of 1908 involved the publication in Germany of an interview with a British daily newspaper that included wild statements and diplomatically damaging remarks. Wilhelm had seen the interview as an opportunity to promote his views and ideas on Anglo-German friendship, but due to his emotional outbursts during the course of the interview, he ended up further alienating not only the British, but also the French, Russians, and Japanese. He implied, among other things, that the Germans cared nothing for the British; that the French and Russians had attempted to incite Germany to intervene in the Second Boer War; and that the German naval buildup was targeted against the Japanese, not Britain. One memorable quotation from the interview was, "You English are mad, mad, mad as March hares." The effect in Germany was quite significant, with serious calls for his abdication. Wilhelm kept a very low profile for many months after the Daily Telegraph fiasco, but later exacted his revenge by forcing the resignation of the chancellor, Prince Bülow, who had abandoned the Emperor to public scorn by not having the transcript edited before its German publication. The Daily Telegraph crisis deeply wounded Wilhelm's previously unimpaired self-confidence, and he soon suffered a severe bout of depression from which he never fully recovered. He lost much of the influence he had previously exercised in domestic and foreign policy.
Naval arms race with Britain
Nothing Wilhelm did in the international arena was of more influence than his decision to pursue a policy of massive naval construction. A powerful navy was Wilhelm's pet project. He had inherited from his mother a love of the British Royal Navy, which was at that time the world's largest. He once confided to his uncle, the Prince of Wales, that his dream was to have a "fleet of my own some day". Wilhelm's frustration over his fleet's poor showing at the Fleet Review at his grandmother Queen Victoria's Diamond Jubilee celebrations, combined with his inability to exert German influence in South Africa following the dispatch of the Kruger telegram, led to Wilhelm taking definitive steps toward the construction of a fleet to rival that of his British cousins. Wilhelm called on the services of the dynamic naval officer Alfred von Tirpitz, whom he appointed to the head of the Imperial Naval Office in 1897.
The new admiral had conceived of what came to be known as the "Risk Theory" or the Tirpitz Plan, by which Germany could force Britain to accede to German demands in the international arena through the threat posed by a powerful battlefleet concentrated in the North Sea. Tirpitz enjoyed Wilhelm's full support in his advocacy of successive naval bills of 1897 and 1900, by which the German navy was built up to contend with that of the British Empire. Naval expansion under the Fleet Acts eventually led to severe financial strains in Germany by 1914, as by 1906 Wilhelm had committed his navy to construction of the much larger, more expensive dreadnought type of battleship.
In 1889 Wilhelm reorganised top-level control of the navy by creating a Naval Cabinet (Marine-Kabinett) equivalent to the German Imperial Military Cabinet which had previously functioned in the same capacity for both the army and navy. The Head of the Naval Cabinet was responsible for promotions, appointments, administration, and issuing orders to naval forces. Captain Gustav von Senden-Bibran was appointed as the first head and remained so until 1906. The existing Imperial admiralty was abolished, and its responsibilities divided between two organisations. A new position was created, equivalent to the supreme commander of the army: the Chief of the High Command of the Admiralty, or Oberkommando der Marine, was responsible for ship deployments, strategy and tactics. Vice-Admiral Max von der Goltz was appointed in 1889 and remained in post until 1895. Construction and maintenance of ships and obtaining supplies was the responsibility of the State Secretary of the Imperial Navy Office (Reichsmarineamt), responsible to the Imperial Chancellor and advising the Reichstag on naval matters. The first appointee was Rear Admiral Karl Eduard Heusner, followed shortly by Rear Admiral Friedrich von Hollmann from 1890 to 1897. Each of these three heads of department reported separately to Wilhelm.
In addition to the expansion of the fleet, the Kiel Canal was opened in 1895, enabling faster movements between the North Sea and the Baltic Sea.
World War I
Historians typically argue that Wilhelm was largely confined to ceremonial duties during the war—there were innumerable parades to review and honours to award. "The man who in peace had believed himself omnipotent became in war a 'shadow Kaiser', out of sight, neglected, and relegated to the sidelines."
The Sarajevo crisis
Wilhelm was a friend of Archduke Franz Ferdinand of Austria, and he was deeply shocked by his assassination on 28 June 1914. Wilhelm offered to support Austria-Hungary in crushing the Black Hand, the secret organisation that had plotted the killing, and even sanctioned the use of force by Austria against the perceived source of the movement—Serbia (this is often called "the blank cheque"). He wanted to remain in Berlin until the crisis was resolved, but his courtiers persuaded him instead to go on his annual cruise of the North Sea on 6 July 1914. Wilhelm made erratic attempts to stay on top of the crisis via telegram, and when the Austro-Hungarian ultimatum was delivered to Serbia, he hurried back to Berlin. He reached Berlin on 28 July, read a copy of the Serbian reply, and wrote on it:
Unknown to the Emperor, Austro-Hungarian ministers and generals had already convinced the 83-year-old Franz Joseph I of Austria to sign a declaration of war against Serbia. As a direct consequence, Russia began a general mobilisation to attack Austria in defence of Serbia.
July 1914
On the night of 30 July 1914, when handed a document stating that Russia would not cancel its mobilisation, Wilhelm wrote a lengthy commentary containing these observations:
More recent British authors state that Wilhelm II really declared, "Ruthlessness and weakness will start the most terrifying war of the world, whose purpose is to destroy Germany. Because there can no longer be any doubts, England, France and Russia have conspired themselves together to fight an annihilation war against us".
When it became clear that Germany would experience a war on two fronts and that Britain would enter the war if Germany attacked France through neutral Belgium, the panic-stricken Wilhelm attempted to redirect the main attack against Russia. When Helmuth von Moltke (the younger) (who had chosen the old plan from 1905, made by General von Schlieffen for the possibility of German war on two fronts) told him that this was impossible, Wilhelm said: "Your uncle would have given me a different answer!" Wilhelm is also reported to have said, "To think that George and Nicky should have played me false! If my grandmother had been alive, she would never have allowed it." In the original Schlieffen Plan, Germany would attack the (supposed) weaker enemy first, meaning France. The plan supposed that it would take a long time before Russia was ready for war. Defeating France had been easy for Prussia in the Franco-Prussian War in 1870. At the 1914 border between France and Germany, an attack at this more southern part of France could be stopped by the French fortress along the border. However, Wilhelm II stopped any invasion of the Netherlands.
Shadow-Kaiser
Wilhelm's role in wartime was one of ever-decreasing power as he increasingly handled awards ceremonies and honorific duties. The high command continued with its strategy even when it was clear that the Schlieffen plan had failed. By 1916 the Empire had effectively become a military dictatorship under the control of Field Marshal Paul von Hindenburg and General Erich Ludendorff. Increasingly cut off from reality and the political decision-making process, Wilhelm vacillated between defeatism and dreams of victory, depending upon the fortunes of his armies. Nevertheless, Wilhelm still retained the ultimate authority in matters of political appointment, and it was only after his consent had been gained that major changes to the high command could be effected. Wilhelm was in favour of the dismissal of Helmuth von Moltke the Younger in September 1914 and his replacement by Erich von Falkenhayn. In 1917, Hindenburg and Ludendorff decided that Bethman-Hollweg was no longer acceptable to them as Chancellor and called upon the Kaiser to appoint somebody else. When asked whom they would accept, Ludendorff recommended Georg Michaelis, a nonentity whom he barely knew. Despite this, the Kaiser accepted the suggestion. Upon hearing in July 1917 that his cousin George V had changed the name of the British royal house to Windsor, Wilhelm remarked that he planned to see Shakespeare's play The Merry Wives of Saxe-Coburg-Gotha. The Kaiser's support base collapsed completely in October–November 1918 in the military, the civilian government, and in German public opinion, as President Woodrow Wilson made it very clear that the monarchy must be overthrown before an end of the war could take place. That year also saw Wilhelm sickened during the worldwide Spanish flu outbreak, though he survived.
Abdication and flight
Wilhelm was at the Imperial Army headquarters in Spa, Belgium, when the uprisings in Berlin and other centres took him by surprise in late 1918. Mutiny among the ranks of his beloved Kaiserliche Marine, the imperial navy, profoundly shocked him. After the outbreak of the German Revolution, Wilhelm could not make up his mind whether or not to abdicate. Up to that point, he accepted that he would likely have to give up the imperial crown, and still hoped to retain the Prussian kingship. However, this was impossible under the imperial constitution. Wilhelm thought he ruled as emperor in a personal union with Prussia. In truth, the constitution defined the empire as a confederation of states under the permanent presidency of Prussia. The imperial crown was thus tied to the Prussian crown, meaning that Wilhelm could not renounce one crown without renouncing the other.
Wilhelm's hope of retaining at least one of his crowns was revealed as unrealistic when, in the hope of preserving the monarchy in the face of growing revolutionary unrest, Chancellor Prince Max of Baden announced Wilhelm's abdication of both titles on 9 November 1918. Prince Max himself was forced to resign later the same day, when it became clear that only Friedrich Ebert, leader of the SPD, could effectively exert control. Later that day, one of Ebert's secretaries of state (ministers), Social Democrat Philipp Scheidemann, proclaimed Germany a republic.
Wilhelm consented to the abdication only after Ludendorff's replacement, General Wilhelm Groener, had informed him that the officers and men of the army would march back in good order under Hindenburg's command, but would certainly not fight for Wilhelm's throne. The monarchy's last and strongest support had been broken, and finally even Hindenburg, himself a lifelong monarchist, was obliged, after polling his generals, to advise the Emperor to give up the crown. On 10 November, Wilhelm crossed the border by train and went into exile in the Netherlands, a neutral. Upon the conclusion of the Treaty of Versailles in early 1919, Article 227 expressly provided for the prosecution of Wilhelm "for a supreme offence against international morality and the sanctity of treaties", but the Dutch government refused to extradite him. King George V wrote that he looked on his cousin as "the greatest criminal in history", but opposed Prime Minister David Lloyd George's proposal to "hang the Kaiser". There was little zeal in Britain to prosecute. On 1 January 1920, it was stated in official circles in London that Great Britain would "welcome refusal by Holland to deliver the former kaiser for trial," and it was hinted that this had been conveyed to the Dutch government through diplomatic channels.
"Punishment of the former kaiser and other German war criminals is worrying Great Britain little, it was said. As a matter of form, however, the British and French governments were expected to request Holland for the former kaiser's extradition. Holland, it was said, will refuse on the ground of constitutional provisions covering the case and then the matter will be dropped. The request for extradition will not be based on genuine desire on the part of British officials to bring the kaiser to trial, according to authoritative information, but is considered necessary formality to 'save the face' of politicians who promised to see that Wilhelm was punished for his crimes."
President Woodrow Wilson of the United States opposed extradition, arguing that prosecuting Wilhelm would destabilise international order and lose the peace.
Wilhelm first settled in Amerongen, where on 28 November he issued a belated statement of abdication from both the Prussian and imperial thrones, thus formally ending the Hohenzollerns' 500-year rule over Prussia. Accepting the reality that he had lost both of his crowns for good, he gave up his rights to "the throne of Prussia and to the German Imperial throne connected therewith." He also released his soldiers and officials in both Prussia and the empire from their oath of loyalty to him. He purchased a country house in the municipality of Doorn, known as Huis Doorn, and moved in on 15 May 1920. This was to be his home for the remainder of his life. The Weimar Republic allowed Wilhelm to remove twenty-three railway wagons of furniture, twenty-seven containing packages of all sorts, one bearing a car and another a boat, from the New Palace at Potsdam.
Life in exile
In 1922, Wilhelm published the first volume of his memoirs—a very slim volume that insisted he was not guilty of initiating the Great War, and defended his conduct throughout his reign, especially in matters of foreign policy. For the remaining twenty years of his life, he entertained guests (often of some standing) and kept himself updated on events in Europe. He grew a beard and allowed his famous moustache to droop, adopting a style very similar to that of his cousins King George V and Tsar Nicholas II. He also learned the Dutch language. Wilhelm developed a penchant for archaeology while residing at the Corfu Achilleion, excavating at the site of the Temple of Artemis in Corfu, a passion he retained in his exile. He had bought the former Greek residence of Empress Elisabeth after her murder in 1898. He also sketched plans for grand buildings and battleships when he was bored. In exile, one of Wilhelm's greatest passions was hunting, and he killed thousands of animals, both beast and bird. Much of his time was spent chopping wood and thousands of trees were chopped down during his stay at Doorn.
Wealth
Wilhelm II was seen as the richest man in Germany before 1914. After his abdication he retained substantial wealth. It was reported that at least 60 railway wagons were needed to carry his furniture, art, porcelain and silver from Germany to the Netherlands. The kaiser retained substantial cash reserves as well as several palaces. After 1945, the Hohenzollerns' forests, farms, factories and palaces in what became East Germany were expropriated and thousands of artworks were subsumed into state-owned museums.
Views on Nazism
In the early 1930s, Wilhelm apparently hoped that the successes of the German Nazi Party would stimulate interest in a restoration of the monarchy, with his eldest grandson as the new Kaiser. His second wife, Hermine, actively petitioned the Nazi government on her husband's behalf. However, Adolf Hitler, himself a veteran of the Imperial German Army during the First World War, felt nothing but contempt for the man he blamed for Germany's greatest defeat, and the petitions were ignored. Though he played host to Hermann Göring at Doorn on at least one occasion, Wilhelm grew to distrust Hitler. Hearing of the murder of the wife of former Chancellor Schleicher during the Night of the Long Knives, Wilhelm said, "We have ceased to live under the rule of law and everyone must be prepared for the possibility that the Nazis will push their way in and put them up against the wall!"
Wilhelm was also appalled at the Kristallnacht of 9–10 November 1938, saying "I have just made my views clear to Auwi [August Wilhelm, Wilhelm's fourth son] in the presence of his brothers. He had the nerve to say that he agreed with the Jewish pogroms and understood why they had come about. When I told him that any decent man would describe these actions as gangsterisms, he appeared totally indifferent. He is completely lost to our family". Wilhelm also stated, "For the first time, I am ashamed to be a German."
In the wake of the German victory over Poland in September 1939, Wilhelm's adjutant, , wrote on his behalf to Hitler, stating that the House of Hohenzollern "remained loyal" and noted that nine Prussian Princes (one son and eight grandchildren) were stationed at the front, concluding "because of the special circumstances that require residence in a neutral foreign country, His Majesty must personally decline to make the aforementioned comment. The Emperor has therefore charged me with making a communication." Wilhelm greatly admired the success which Hitler was able to achieve in the opening months of the Second World War, and personally sent a congratulatory telegram when the Netherlands surrendered in May 1940: "My Fuhrer, I congratulate you and hope that under your marvellous leadership the German monarchy will be restored completely." Unimpressed, Hitler remarked to Linge, his valet, "What an idiot!" Upon the fall of Paris a month later, Wilhelm sent another telegram: "Under the deeply moving impression of France's capitulation I congratulate you and all the German armed forces on the God-given prodigious victory with the words of Kaiser Wilhelm the Great of the year 1870: "What a turn of events through God's dispensation!" All German hearts are filled with the chorale of Leuthen, which the victors of Leuthen, the soldiers of the Great King sang: Now thank we all our God!" Hitler's delayed response was reportedly uninspired and did not reciprocate the former Emperor's enthusiasm. In a letter to his daughter Victoria Louise, Duchess of Brunswick, he wrote triumphantly, "Thus is the pernicious Entente Cordiale of Uncle Edward VII brought to nought." In a September 1940 letter to an American journalist, Wilhelm praised Hitler's rapid early conquests as "a succession of miracles", but remarked also that "the brilliant leading Generals in this war came from My school, they fought under my command in the World War as lieutenants, captains and young majors. Educated by Schlieffen they put the plans he had worked out under me into practice along the same lines as we did in 1914." After the German conquest of the Netherlands in 1940, the aging Wilhelm retired completely from public life. In May 1940, Wilhelm declined an offer from Winston Churchill of asylum in Britain, preferring to die at Huis Doorn.
Anti-England, anti-Semitic, and anti-Freemason views
During his last year at Doorn, Wilhelm believed that Germany was still the land of monarchy and Christianity, while England was the land of Classical liberalism and therefore of Satan and the Antichrist. He argued that the English nobility were "Freemasons thoroughly infected by Juda". Wilhelm asserted that the "British people must be liberated from Antichrist Juda. We must drive Juda out of England just as he has been chased out of the Continent."
He also believed that the Freemasons and the Jews had caused both world wars, and were aiming for a world empire financed by British and American gold, but that "Juda's plan has been smashed to pieces and they themselves swept out of the European Continent!" Continental Europe was now, Wilhelm wrote, "consolidating and closing itself off from British influences after the elimination of the British and the Jews!" The end result would be a "U.S. of Europe!" In a 1940 letter to his sister Princess Margaret, Wilhelm wrote: "The hand of God is creating a new world & working miracles... We are becoming the U.S. of Europe under German leadership, a united European Continent." He added: "The Jews [are] being thrust out of their nefarious positions in all countries, whom they have driven to hostility for centuries."
Also, in 1940 came what would have been his mother's 100th birthday. Despite their very troubled relationship, Wilhelm wrote to a friend, "Today the 100th birthday of my mother! No notice is taken of it at home! No 'Memorial Service' or ... committee to remember her marvellous work for the ... welfare of our German people ... Nobody of the new generation knows anything about her."
Death
Wilhelm died of a pulmonary embolism in Doorn, Netherlands, on 4 June 1941, at the age of 82, just weeks before the Axis invasion of the Soviet Union. Despite his personal animosity toward the monarchy, Hitler wanted to bring the Kaiser's body back to Berlin for a state funeral, as Hitler felt that such a funeral, with himself acting in the role of heir apparent to the throne, would be useful to exploit for propaganda. However, Wilhelm's orders that his body was not to return to Germany unless the monarchy was first restored were then revealed and were grudgingly respected. The Nazi occupation authorities arranged for a small military funeral, with only a few hundred people present. The mourners included Field Marshal August von Mackensen, fully dressed in his old Imperial Hussars uniform, Admiral Wilhelm Canaris, Colonel General Curt Haase, World War I flying ace turned Wehrmachtbefehlshaber for the Netherlands General Friedrich Christiansen, and Reichskommissar for the Netherlands Arthur Seyss-Inquart, along with a few other military advisers. However, Wilhelm's insistence that the swastika and Nazi Party regalia not be displayed at his funeral was ignored, as may accordingly be seen in the photographs of the funeral taken by a Dutch photographer.
Wilhelm was buried in a mausoleum upon the grounds of Huis Doorn, which has since become a place of pilgrimage for German monarchists, who gather there every year on the anniversary of his death to pay their homage to the last German Emperor.
Historiography
Three trends have characterised the writing about Wilhelm. First, the court-inspired writers considered him a martyr and a hero, often uncritically accepting the justifications provided in the Kaiser's own memoirs. Second, there came those who judged Wilhelm to be completely unable to handle the great responsibilities of his position, a ruler too reckless to deal with power. Third, after 1950, later scholars have sought to transcend the passions of the early 20th century and attempted an objective portrayal of Wilhelm and his rule.
On 8 June 1913, a year before the Great War began, The New York Times published a special supplement devoted to the 25th anniversary of the Kaiser's accession. The banner headline read: "Kaiser, 25 Years a Ruler, Hailed as Chief Peacemaker". The accompanying story called him "the greatest factor for peace that our time can show", and credited Wilhelm with frequently rescuing Europe from the brink of war. Until the late 1950s, Germany under the last Kaiser was depicted by most historians as an almost absolute monarchy. Partly, however, this was a deliberate deception by German civil servants and elected officials. For example, President Theodore Roosevelt believed the Kaiser was in control of German foreign policy because Hermann Speck von Sternburg, the German ambassador in Washington and a personal friend of Roosevelt, presented to the President messages from Chancellor von Bülow as though they were messages from the Kaiser. Later historians downplayed his role, arguing that senior officials regularly learned to work around the Kaiser's back. More recently, historian John C. G. Röhl has portrayed Wilhelm as the key figure in understanding the recklessness and downfall of Imperial Germany. Thus, the argument is still made that the Kaiser played a major role in promoting the policies of both naval and colonialist expansion that caused the deterioration of Germany's relations with Britain before 1914.
Marriages and issue
Wilhelm and his first wife, Princess Augusta Victoria of Schleswig-Holstein, were married on 27 February 1881. They had seven children:
Empress Augusta, known affectionately as "Dona", was a constant companion to Wilhelm, and her death on 11 April 1921 was a devastating blow. It also came less than a year after their son Joachim committed suicide.
Remarriage
The following January, Wilhelm received a birthday greeting from a son of the late Prince Johann George Ludwig Ferdinand August Wilhelm of Schönaich-Carolath. The 63-year-old Wilhelm invited the boy and his mother, Princess Hermine Reuss of Greiz, to Doorn. Wilhelm found Hermine very attractive, and greatly enjoyed her company. The couple were wed in Doorn on 5 November 1922 despite the objections of Wilhelm's monarchist supporters and his children. Hermine's daughter, Princess Henriette, married the late Prince Joachim's son, Karl Franz Josef, in 1940, but divorced in 1946. Hermine remained a constant companion to the aging former emperor until his death.
Religion
Own views
In accordance with his role as the King of Prussia, Emperor Wilhelm II was a Lutheran member of the Evangelical State Church of Prussia's older Provinces. It was a United Protestant denomination, bringing together Reformed and Lutheran believers.
Attitude towards Islam
Wilhelm II was on friendly terms with the Muslim world. He described himself as a "friend" to "300 million Mohammedans". Following his trip to Constantinople (which he visited three times—an unbeaten record for any European monarch) in 1898, Wilhelm II wrote to Nicholas II that,
in response to the political competition between the Christian sects to build bigger and grander churches and monuments which made the sects appear idolatrous and turned Muslims away from the Christian message.
Antisemitism
Wilhelm's biographer Lamar Cecil identified Wilhelm's "curious but well-developed anti-Semitism", noting that in 1888 a friend of Wilhelm "declared that the young Kaiser's dislike of his Hebrew subjects, one rooted in a perception that they possessed an overweening influence in Germany, was so strong that it could not be overcome".
Cecil concludes:
Wilhelm never changed, and throughout his life he believed that Jews were perversely responsible, largely through their prominence in the Berlin press and in leftist political movements, for encouraging opposition to his rule. For individual Jews, ranging from rich businessmen and major art collectors to purveyors of elegant goods in Berlin stores, he had considerable esteem, but he prevented Jewish citizens from having careers in the army and the diplomatic corps and frequently used abusive language against them.
In 1918, Wilhelm suggested a campaign against the "Jew-Bolsheviks" in the Baltic states, citing the example of what Turks had done to the Armenians a few years earlier.
On 2 December 1919, Wilhelm wrote to Field Marshal August von Mackensen, denouncing his own abdication as the "deepest, most disgusting shame ever perpetrated by a person in history, the Germans have done to themselves ... egged on and misled by the tribe of Judah ... Let no German ever forget this, nor rest until these parasites have been destroyed and exterminated from German soil!" Wilhelm advocated a "regular international all-worlds pogrom à la Russe" as "the best cure" and further believed that Jews were a "nuisance that humanity must get rid of some way or other. I believe the best thing would be gas!"
Documentaries and films
William II. – The last days of the German Monarchy (original title: "Wilhelm II. – Die letzten Tage des Deutschen Kaiserreichs"), about the abdication and flight of the last German Kaiser. Germany/Belgium, 2007. Produced by seelmannfilm and German Television. Written and directed by Christoph Weinert.
Queen Victoria and the Crippled Kaiser, Channel 4, Secret History series 13; first broadcast 17 November 2013
Barry Foster played the adult Wilhelm II in several episodes of the 1974 BBC TV series Fall of Eagles.
Christopher Neame played Wilhelm II in the several episodes of the 1975 BBC TV series Edward the Seventh.
Rupert Julian played Wilhelm II in the 1918 Hollywood propaganda film The Kaiser, the Beast of Berlin.
Alfred Struwe played Wilhelm in the 1987 Polish historical drama film Magnat.
Robert Stadlober played a young crown prince Wilhelm and friend of Rudolf, Crown Prince of Austria in the acclaimed 2006 film Kronprinz Rudolf (The Crown Prince).
Ladislav Frej played the Kaiser in the 2008 film The Red Baron.
Rainer Sellien played Wilhelm II in the 2014 BBC miniseries 37 Days.
Christopher Plummer played a depressed Wilhelm II living in exile at Huis Doorn in the 2016 romantic war drama The Exception.
Tom Hollander played Wilhelm II in the 2021 movie The King's Man.
Orders and decorations
German honours
Foreign honours
Ancestry
See also
Ålesund, a Norwegian city rebuilt by Wilhelm II after it had been almost completely destroyed by fire in 1904
German entry into World War I
Rulers of Germany family tree
Wilhelminism on society, politics, culture, art and architecture of Germany 1890–1918
References
Works cited
.
. online
.
, Archive.org.
.
.
.
, translated in
.
.
.
"The interview of the Emperor Wilhelm II on October 28, 1908" London Daily Telegraph Online
Further reading
Clark, Christopher M. Kaiser Wilhelm II. (2000) 271 pp. short biography by scholar
Domeier, Norman. The Eulenburg Affair: A Cultural History of Politics in the German Empire (2015)
Eley, Geoff. "The View From The Throne: The Personal Rule of Kaiser Wilhelm II," Historical Journal, June 1985, Vol. 28 Issue 2, pp. 469–485.
Haardt, Oliver FR. "The Kaiser in the Federal State, 1871–1918." German History 34.4 (2016): 529–554. online
Kohut, Thomas A. Wilhelm II and the Germans: A Study in Leadership, New York: Oxford University Press, 1991. .
Langer, William L. The Diplomacy of Imperialism, 1890–1902 (1935)online
Mommsen, Wolfgang J. "Kaiser Wilhelm II and German Politics." Journal of Contemporary History 1990 25(2–3): 289–316. argues his irrationality and instability made worse the weaknesses in Germany's constitutional and political systems
Otte, T.G., "'The Winston of Germany': The British Elite and the Last German Emperor" Canadian Journal of History 36 (December 2001). They though he was mental unstable and this helped shape British policy.
Retallack, James. Germany in the Age of Kaiser Wilhelm II (St. Martin's Press, 1996). .
Rich, Norman. "The Question of National Interest In Imperial German Foreign Policy: Bismarck, William II, and the Road to World War I." Naval War College Review (1973) 26#1: 28–41. online
Röhl, John C. G; Sombart, Nicolaus, eds. Kaiser Wilhelm II New Interpretations: The Corfu Papers, (Cambridge UP, 1982)
Van der Kiste, John. Kaiser Wilhelm II: Germany's Last Emperor, Sutton Publishing, 1999. .
Waite, Robert GL Kaiser and Führer: A Comparative Study of Personality and Politics (1998) 511 pp. Psychohistory comparing him to Adolf Hitler.
External links
The German Emperor as shown in his public utterances
, Google Books.
The German emperor's speeches: being a selection from the speeches, edicts, letters, and telegrams of the Emperor William II
, mostly in German
Historical film documents on Wilhelm II from the time of World War I at European Film Gateway
The 1922 book review of My Memoir from The Spectator
1859 births
1941 deaths
19th-century German emperors
20th-century German emperors
19th-century Kings of Prussia
19th-century German people
Anti-English sentiment
Anti-Judaism
Anti-Masonry
Antisemitism in Germany
Burials in Utrecht (province)
Crown Princes of Prussia
Deaths from pulmonary embolism
Dethroned monarchs
German anti-communists
German conspiracy theorists
German emperors
German exiles
German expatriates in the Netherlands
German Lutherans
German nationalists
German people of English descent
German politicians with physical disabilities
German racehorse owners and breeders
House of Hohenzollern
Kings of Prussia
Monarchs who abdicated
People from Berlin
People from Doorn
Pretenders to the German throne
Protestant monarchs
Royalty and nobility with disabilities
Spanish captain generals
University of Bonn alumni
Sons of emperors
Recipients of the Grand Cross of the Iron Cross
Recipients of the Pour le Mérite (military class)
Grand Crosses of the Military Order of Max Joseph
Recipients of the Hanseatic Cross (Bremen)
Recipients of the Hanseatic Cross (Lübeck)
Recipients of the Military Merit Cross (Mecklenburg-Schwerin), 1st class
Grand Crosses of the Order of Saint Stephen of Hungary
Grand Crosses of the Military Order of Maria Theresa
Recipients of the Order of Bravery, 1st class
Recipients of the Cross of Honour of the Order of the Dannebrog
Knights Grand Cross of the Order of Saints Maurice and Lazarus
Knights Grand Cross of the Military Order of Savoy
Recipients of the Order of the Netherlands Lion
Knights Grand Cross of the Military Order of William
Grand Crosses of the Order of the House of Orange
Knights of the Order of the Norwegian Lion
2
2
Recipients of the Order of the White Eagle (Russia)
Recipients of the Order of St. Anna, 1st class
Recipients of the Order of Saint Stanislaus (Russian), 1st class
Knights of the Golden Fleece of Spain
Grand Crosses of the Order of Vasa
Annulled Honorary Knights Grand Cross of the Royal Victorian Order
Extra Knights Companion of the Garter
Knights of Justice of the Order of St John
People stripped of honorary degrees |
null | null | Inkjet printing | eng_Latn | Inkjet printing is a type of computer printing that recreates a digital image by propelling droplets of ink onto paper and plastic substrates. Inkjet printers were the most commonly used type of printer in 2008, and range from small inexpensive consumer models to expensive professional machines. By 2019, laser printers outsold inkjet printers by nearly a 2:1 ratio, 9.6% vs 5.1%.
The concept of inkjet printing originated in the 20th century, and the technology was first extensively developed in the early 1950s. While working at Canon in Japan, Ichiro Endo suggested the idea for a "Bubble jet" printer, while around the same time Jon Vaught at HP was developing a similar idea. In the late 1970s, inkjet printers that could reproduce digital images generated by computers were developed, mainly by Epson, Hewlett-Packard (HP) and Canon. In the worldwide consumer market, four manufacturers account for the majority of inkjet printer sales: Canon, HP, Epson and Brother.
In 1982, Robert Howard came up with the idea to produce a small color printing system that used piezos to spit drops of ink. He formed the company, R.H. (Robert Howard) Research (named Howtek, Inc. in Feb 1984), and developed the revolutionary technology that led to the Pixelmaster color printer with solid ink using Thermojet technology. This technology consists of a tubular single nozzle acoustical wave drop generator invented originally by Steven Zoltan in 1972 with a glass nozzle and improved by the Howtek inkjet engineer in 1984 with a Tefzel molded nozzle to remove unwanted fluid frequencies.
The emerging ink jet material deposition market also uses inkjet technologies, typically printheads using piezoelectric crystals, to deposit materials directly on substrates.
The technology has been extended and the 'ink' can now also comprise solder paste in PCB assembly, or living cells, for creating biosensors and for tissue engineering.
Images produced on inkjet printers are sometimes sold under trade names such as Digigraph, Iris prints, giclée, and Cromalin. Inkjet-printed fine art reproductions are commonly sold under such trade names to imply a higher quality product and avoid association with everyday printing.
Methods
Fluid surface tension naturally pulls a stream into droplets. Optimal drop sizes of 0.004 inch require an inkjet nozzle size of about 0.003 inches. Fluids with surface tension may be water based, wax or oil based and even melted metal alloys. Most drops can be electrically charged. There are two main technologies in use in contemporary inkjet printers: continuous (CIJ) and drop-on-demand (DOD). Continuous inkjet means the flow is pressurized and in a continuous stream. Drop-on-demand means the fluid is expelled from the jet nozzle one drop at a time. This can be done with a mechanical means with a push or some electrical method. A large electrical charge can pull drops out of a nozzle, sound waves can push fluid from a nozzle or a chamber volume expansion can expel a drop. Continuous streaming was investigated first many years ago. Drop-on-demand was only discovered in the 1920s.
Continuous inkjet
The continuous inkjet (CIJ) method is used commercially for marking and coding of products and packages. In 1867, Lord Kelvin patented the syphon recorder, which recorded telegraph signals as a continuous trace on paper using an ink jet nozzle deflected by a magnetic coil. The first commercial devices (medical strip chart recorders) were introduced in 1951 by Siemens. using the patent US2566443 invented by Elmquist Rune dated 9/4/1951.
In CIJ technology, a high-pressure pump directs liquid ink from a reservoir through a gunbody and a microscopic nozzle (usually .003 inch diameter), creating a continuous stream of ink droplets via the Plateau-Rayleigh instability. A piezoelectric crystal may be used to create an acoustic wave as it vibrates within the gunbody and causes the stream of liquid to break into droplets at regular intervals: 64,000 to 165,000 irregular-sized ink droplets per second may be achieved. The ink droplets are subjected to an electrostatic field created by a charging electrode or by a magnetic flux field as they form; the field varies according to the degree of drop deflection desired. This results in a controlled deflection by electrostatic charge on each droplet. Charged droplets may be separated by one or more uncharged "guard droplets" to minimize electrostatic repulsion between neighboring droplets.
The droplets pass through another electrostatic or magnetic field and are directed (deflected) by electrostatic deflection plates or flux field to print on the receptor material (substrate), or allowed to continue on deflected to a collection gutter for re-use. The more highly charged droplets are deflected to a greater degree. Only a small fraction of the droplets is used to print, the majority being recycled.
CIJ is one of the oldest (1951) ink jet technologies in use and is fairly mature. Drop-on-demand was not invented until later. The major advantages of CIJ are the very high velocity (≈20 m/s) of the ink droplets, which allows for a relatively long distance between print head and substrate, and the very high drop ejection frequency, allowing for very high speed printing. Another advantage is freedom from nozzle clogging as the jet is always in use, therefore allowing volatile solvents such as ketones and alcohols to be employed, giving the ink the ability to "bite" into the substrate and dry quickly. The ink system requires active solvent regulation to counter solvent evaporation during the time of flight (time between nozzle ejection and gutter recycling), and from the venting process whereby air that is drawn into the gutter along with the unused drops is vented from the reservoir. Viscosity is monitored and a solvent (or solvent blend) is added to counteract solvent loss.
In the later 1950s heated wax inks became popular with CIJ technologies. In 1971 Johannes F. Gottwald patent US3596285A, Liquid Metal Recorder used molten metal ink with a magnetic flux field to fabricate formed symbols for signage. This may have been the first 3D metal object printed using magnetic core memory as data to produce each symbol.
Drop-on-demand
There are many ways to produce a drop-on-demand (DOD) inkjet. Common methods include thermal DOD and piezoelectric DOD to speed up the frequency of drops. DOD may use a single nozzle or thousands of nozzles. One DOD process uses software that directs the heads to apply between zero and eight droplets of ink per dot, only where needed. Inkjet fluid materials have expanded to include pastes, epoxies, hot-melt inks, biological fluids, etc. DOD is very popular and has an interesting history. Mechanical DOD came first, followed by electrical methods including piezoelectric devices and then thermal or heat expansion methods.
Thermal DOD printing Most consumer inkjet printers, including those from Canon (FINE Cartridge system, see photo), Hewlett-Packard, and Lexmark, use the thermal inkjet process. The idea of using thermal excitation to move tiny drops of ink was developed independently by two groups at roughly the same time: John Vaught and a team at Hewlett-Packard's Corvallis Division, and Canon engineer Ichiro Endo. Initially, in 1977, Endo's team was trying to use the piezoelectric effect to move ink out of the nozzle but noticed that ink shot out of a syringe when it was accidentally heated with a soldering iron. Vaught's work started in late 1978 with a project to develop fast, low-cost printing. The team at HP found that thin-film resistors could produce enough heat to fire an ink droplet. Two years later the HP and Canon teams found out about each other's work.
Thermal inkjet In the thermal inkjet process, the print cartridges consist of a series of tiny chambers, each containing a heater, all of which are constructed by photolithography. To eject a droplet from each chamber, a pulse of current is passed through the heating element causing a rapid vaporization of the ink in the chamber and forming a bubble, which causes a large pressure increase, propelling a droplet of ink onto the paper (hence Canon's trade name of Bubble Jet). Early thermal heads ran at just 600-700 dpi but improvements by HP increased the firing range of 8-12 khz per chamber and as high as 18khz with 5-picoliter drop volume by the year 2000. Thermal printheads do not have the power of piezo DOD or continuous inkjet, so the gap between the face of the head and paper is critical. The ink's surface tension, as well as the condensation and resultant contraction of the vapor bubble, pulls a further charge of ink into the chamber through a narrow channel attached to an ink reservoir. The inks involved are usually water-based and use either pigments or dyes as the colorant. The inks must have a volatile component to form the vapor bubble; otherwise droplet ejection cannot occur. As no special materials are required, the print head is generally cheaper to produce than in other inkjet technologies.
Piezoelectric DOD printing Piezos are electrically polarized ceramic devices, just as a magnet is polarized. Most commercial and industrial inkjet printers and some consumer printers (those produced by Epson (see photo) and Brother Industries) use a piezoelectric material in an ink-filled chamber behind each nozzle instead of a heating element. When a voltage is applied, the piezoelectric material changes shape, generating a pressure pulse in the fluid, which pushes a droplet of ink from the nozzle. Single nozzle tubular inkjets actually are fluid resonator chambers and the drops are expelled by sound waves in the ink chamber. The 1972 patent called them squeeze tube inkjets but later it was discovered to be acoustical inkjets. Piezoelectric (also called piezo) inkjet allows a wider variety of inks than thermal inkjet as there is no requirement for a volatile component, and no issue with kogation (buildup of ink residue), but the print heads are more expensive to manufacture due to the use of piezoelectric material (usually PZT, lead zirconium titanate). However, the ink cartridges can be separate from the head itself and individually be replaced as needed. Piezo, then has the potential for lower running costs. Piezo heads are said to achieve firing rates that are faster than thermal heads at comparable drop volumes.
Piezo inkjet Piezo inkjet technology is often used on production lines to mark products. For instance, the "use-before" date is often applied to products with this technique; in this application the head is stationary and the product moves past. This application requires a relatively large gap between the print head and the substrate, but also yields a high speed, a long service life, and low operating cost.
Thermoplastic/3D printing In the 1970s, the first DOD inks were water-based and higher temperature use was not recommended. In the late 1970s wax and oil based inks were used in the Silonics in 1975, Siemens PT-80i in 1977 and Epson and Exxon in 1980's DOD inkjets. In 1984 a small company, Howtek, Inc., found that solid ink materials (thermoplastics) could be jetted at 125 °C by maintaining the piezo poling charge while printing. In 1986 Howtek launched the Pixelmaster solid ink-jetting printer, which opened the door to printing three-dimensional plastic inks and led to a 1992 3D patent, US5136515A. This patent was licensed by the first 3 major 3D printer companies (Sanders Prototype, Inc, Stratasys and 3D Systems).
Braillemaster In the late 1980s, Howtek introduced the Braillemaster, a printer that used four layers of solid ink per character to create documents in Braille that could be read by people who were blind.
Howtek Solidscape, Inc., currently uses the Howtek-style thermoplastic materials and Howtek-style single nozzle inkjets (see illustration) very successfully. Ballistic Particle Manufacturing also used the Howtek style materials and inkjets. These inkjets can produce up to 16,000 drops per second and shoot drops at 9 feet per second. Originally designed to only print on standard letter-sized paper sheets they now can print 3D models requiring hundreds of layers.
Thermojet The thermoplastic inks in piezoelectric inkjets (called Thermojet technology by Howtek) are sometimes confused with the thermal (heat expansion) bubble-jet technology but they are completely different. Bubble jet inks are not solid at room temp and are not heated. Thermojet inks require 125 °C to reduce fluid viscosity in jetting range. Howtek was the first to introduce an inkjet color printer using thermoplastic inks in 1984 at Comdex, Las Vegas.
Ink formulations
The earliest reference to a continuous inkjet ink (CIJ) in the 1971 patent US3596285A states " The preferred ink is characterized by viscosity and surface tension characteristics such that the liquid will be maintained over span under the force with which it is moving in bridge or stream. Implicit in such requirement is that the pressure applied to the ink in formation of said stream is sufficient to form a jet and to impart enough energy to carry the jet as a continuous liquid mass notwithstanding the defective forces which are or may be applied. Furthermore, the color of the ink and the color of the carrier should be such that good optical contrast is formed there between the following printing. The preferred ink is a "hot-melt ink". That is to say it will assume a solid phase at the temperature of the carrier and liquid phase at some higher temperature. The range of commercially available ink compositions which could meet the requirement of the invention are not known at the present time. However, satisfactory printing according to the invention has been achieved with a conductive metal alloy as ink. It is extremely hard at room temperature and adheres well to the surface of the carrier.
The basic problem with inkjet inks is the conflicting requirements for a coloring agent that will stay on the surface versus rapid bleed through the carrier fluid.
Desktop inkjet printers, as used in offices and homes, tend to use aqueous ink based on a mixture of water, glycol and dyes or pigments. These inks are inexpensive to manufacture but difficult to control on the surface of media, often requiring specially coated media. HP inks contain sulfonated polyazo black dye (commonly used for dyeing leather), nitrates and other compounds. Aqueous inks are mainly used in printers with thermal inkjet heads, as these heads require water to perform the ink-expelling function.
While aqueous inks often provide the broadest color gamut and most vivid color, most are not waterproof without specialized coating or lamination after printing. Most dye-based inks, while usually the least expensive, are subject to rapid fading when exposed to light or ozone. Pigment-based aqueous inks are typically more costly but provide much better long-term durability and ultraviolet resistance. Inks marketed as "archival quality" are usually pigment-based.
Some professional wide format printers use aqueous inks, but the majority in professional use today employ a much wider range of inks, most of which require piezo inkjet heads and extensive maintenance.
Solvent inks The main ingredient of these inks are volatile organic compounds (VOCs), organic chemical compounds that have high vapor pressures. Color is achieved with pigments rather than dyes for excellent fade-resistance. The chief advantage of solvent inks is that they are comparatively inexpensive and enable printing on flexible, uncoated vinyl substrates, which are used to produce vehicle graphics, billboards, banners and adhesive decals. Disadvantages include the vapour produced by the solvent and the need to dispose of used solvent. Unlike most aqueous inks, prints made using solvent-based inks are generally waterproof and ultraviolet-resistant (for outdoor use) without special over-coatings. The high print speed of many solvent printers demands special drying equipment, usually a combination of heaters and blowers. The substrate is usually heated immediately before and after the print heads apply ink. Solvent inks are divided into two sub-categories: hard solvent ink offers the greatest durability without specialized over-coatings but requires specialized ventilation of the printing area to avoid exposure to hazardous fumes, while mild or "eco" solvent inks, while still not as safe as aqueous inks, are intended for use in enclosed spaces without specialized ventilation of the printing area. Mild solvent inks have rapidly gained popularity in recent years as their color quality and durability have increased while ink cost has dropped significantly.
UV-curable inks These inks consist mainly of acrylic monomers with an initiator package. After printing, the ink is cured by exposure to strong UV light. Ink is exposed to UV radiation where a chemical reaction takes place where the photo-initiators cause the ink components to cross-link into a solid. Typically a shuttered mercury-vapor lamp or UV LED is used for the curing process. Curing processes with high power for short periods of times (microseconds) allow curing inks on thermally sensitive substrates. UV inks do not evaporate, but rather cure or set as a result from this chemical reaction. No material is evaporated or removed, which means about 100% of the delivered volume is used to provide coloration. This reaction happens very quickly, which leads to instant drying that results in a completely cured graphic in a matter of seconds. This also allows for a very fast print process. As a result of this instant chemical reaction no solvents penetrate the substrate once it comes off the printer, which allows for high quality prints. The advantage of UV-curable inks is that they "dry" as soon as they are cured, they can be applied to a wide range of uncoated substrates, and they produce a very robust image. Disadvantages are that they are expensive, require expensive curing modules in the printer, and the cured ink has a significant volume and so gives a slight relief on the surface. Though improvements are being made in the technology, UV-curable inks, because of their volume, are somewhat susceptible to cracking if applied to a flexible substrate. As such, they are often used in large "flatbed" printers, which print directly to rigid substrates such as plastic, wood or aluminium where flexibility is not a concern.
Dye sublimation inks These inks contain special sublimation dyes and are used to print directly or indirectly on to fabrics which consist of a high percentage of polyester fibers. A heating step causes the dyes to sublimate into the fibers and create an image with strong color and good durability.
Solid inkThese inks consist mainly of waxy compounds which are heated past their melting point to enable printing, and which harden upon hitting the cooled substrate. Hot-melt inks are typically used for masking processes and are found in graphic printing. The earliest hot-melt ink was patented in 1971 by Johannes F Gottwald, US3596285A, Liquid Metal Recorder was intended for printing. The patent states "As used herein the term "printing" is not intended in a limited sense but includes writing or other symbol or pattern formulation with an ink. The term ink as used is intended to include not only dye or pigment-containing materials, but any flowable substance or composition suited for application to surface for forming symbols, characters or patterns of intelligence by marking. The materials employed in such process can be salvaged for reuse. Another object of the invention is to increase the size of characters.....in terms of material requirements for such large and continuous displays".
Printing heads
There are two main design philosophies in inkjet head design: fixed-head and disposable head. Each has its own strengths and weaknesses.
Fixed head
The fixed-head philosophy provides an inbuilt print head (often referred to as a gaiter-head) that is designed to last for the life of the printer. The idea is that because the head need not be replaced every time the ink runs out, consumable costs can be made lower and the head itself can be more precise than a cheap disposable one, typically requiring no calibration. On the other hand, if a fixed head is damaged, obtaining a replacement head can become expensive, if removing and replacing the head is even possible. If the printer's head cannot be removed, the printer itself will then need to be replaced.
Fixed head designs are available in consumer products, but are more likely to be found on industrial high-end printers and large format plotters. In the consumer space, fixed-head printers are manufactured primarily by Epson and Canon; however, many more recent Hewlett-Packard models use a fixed-head, such as the Officejet Pro 8620 and HP's Pagewide series
Disposable head
The disposable head philosophy uses a print head which is supplied as a part of a replaceable ink cartridge. Every time a cartridge is exhausted, the entire cartridge and print head are replaced with a new one. This adds to the cost of consumables and makes it more difficult to manufacture a high-precision head at a reasonable cost, but also means that a damaged or clogged print head is only a minor problem: the user can simply buy a new cartridge. Hewlett-Packard has traditionally favoured the disposable print head, as did Canon in its early models. This type of construction can also be seen as an effort by printer manufacturers to stem third party ink cartridge assembly replacements, as these would-be suppliers don't have the ability to manufacture specialized print heads.
An intermediate method does exist: a disposable ink tank connected to a disposable head, which is replaced infrequently (perhaps every tenth ink tank or so). Most high-volume Hewlett-Packard inkjet printers use this setup, with the disposable print heads used on lower volume models. A similar approach is used by Kodak, where the printhead intended for permanent use is nevertheless inexpensive and can be replaced by the user. Canon now uses (in most models) replaceable print heads which are designed to last the life of the printer, but can be replaced by the user should they become clogged.
Additive manufacturing 3D printheads have very long operating "print times" and failures will occur from internal clogs, orifice damage from bumping obstacles on the print table, calibration failures from overstressed piezoelectric bond life failures and other unexpected causes. Replacement printheads are on spare parts lists for most long use-life 3D printers.
Cleaning mechanisms
The primary cause of inkjet printing problems is ink drying on the printhead's nozzles, causing the pigments and dyes to dry out and form a solid block of hardened mass that plugs the microscopic ink passageways. Most printers attempt to prevent this drying from occurring by covering the printhead nozzles with a rubber cap when the printer is not in use. Abrupt power losses, or unplugging the printer before it has capped the printhead, can cause the printhead to be left in an uncapped state. Even when the head is capped, this seal is not perfect, and over a period of several weeks the moisture (or other solvent) can still seep out, causing the ink to dry and harden. Once ink begins to collect and harden, the drop volume can be affected, drop trajectory can change, or the nozzle can completely fail to jet ink.
To combat this drying, nearly all inkjet printers include a mechanism to reapply moisture to the printhead. Typically there is no separate supply of pure ink-free solvent available to do this job, and so instead the ink itself is used to remoisten the printhead. The printer attempts to fire all nozzles at once, and as the ink sprays out, some of it wicks across the printhead to the dry channels and partially softens the hardened ink. After spraying, a rubber wiper blade is swept across the printhead to spread the moisture evenly across the printhead, and the jets are again all fired to dislodge any ink clumps blocking the channels.
Some printers use a supplemental air-suction pump, using the rubber capping station to suck ink through a severely clogged cartridge. The suction pump mechanism is frequently driven by the page feed stepper motor: it is connected to the end of the shaft. The pump only engages when the shaft turns backwards, hence the rollers reversing while head cleaning. Due to the built-in head design, the suction pump is also needed to prime the ink channels inside a new printer, and to reprime the channels between ink tank changes.
Professional solvent- and UV-curable ink wide-format inkjet printers generally include a "manual clean" mode that allows the operator to manually clean the print heads and capping mechanism and to replace the wiper blades and other parts used in the automated cleaning processes. The volume of ink used in these printers often leads to "overspray" and therefore buildup of dried ink in many places that automated processes are not capable of cleaning.
The ink consumed in the cleaning process needs to be collected to prevent ink from leaking in the printer. The collection area is called the spittoon, and in Hewlett Packard printers this is an open plastic tray underneath the cleaning/wiping station. In Epson printers, there is typically a large absorption pad in a pan underneath the paper feed platen. For printers several years old, it is common for the dried ink in the spittoon to form a pile that can stack up and touch the printheads, jamming the printer. Some larger professional printers using solvent inks may employ a replaceable plastic receptacle to contain waste ink and solvent which must be emptied or replaced when full.
There is a second type of ink drying that most printers are unable to prevent. For ink to spray from the cartridge, air must enter to displace the removed ink. The air enters via an extremely long, thin labyrinth tube, up to 10 cm long, wrapping back and forth across the ink tank. The channel is long and narrow to reduce moisture evaporation through the vent tube, but some evaporation still occurs and eventually the ink cartridge dries up from the inside out. To combat this problem, which is especially acute with professional fast-drying solvent inks, many wide-format printer cartridge designs contain the ink in an airtight, collapsible bag that requires no vent. The bag merely shrinks until the cartridge is empty.
The frequent cleaning conducted by some printers can consume quite a bit of ink and has a great impact on cost-per-page determinations.
Clogged nozzles can be detected by printing a standard test pattern on the page. Some software workaround methods are known for re-routing printing information from a clogged nozzle to a working nozzle.
Ink delivery developments
Ink cartridges have been the traditional method for delivering ink to the printhead. Continuous ink system (CISS) inkjet printers connect the printhead either to high-capacity ink tanks or packs, or replenish the built-in cartridges via external tanks connected via tubes, typically a retrofit configuration. Supertank printers–a subset of CISS printers–have high-capacity integrated ink tanks or ink packs, and are manually refilled via ink bottles. When supertank ink systems are paired with disposable printhead technology, replaceable cartridges are used to replace the exhausted print heads.
Advantages
Compared to earlier consumer-oriented color printers, inkjet printers have a number of advantages. They are quieter in operation than impact dot matrix or daisywheel printers. They can print finer, smoother details through higher resolution. Consumer inkjet printers with photographic-quality printing are widely available.
In comparison to technologies like thermal wax, dye sublimation, and laser printing, inkjets have the advantage of practically no warm up time, and often lower cost per page. However, low-cost laser printers can have lower per-page costs, at least for black-and-white printing, and possibly for color.
For some inkjet printers, monochrome ink sets are available either from the printer manufacturer or from third-party suppliers. These allow the inkjet printer to compete with the silver-based photographic papers traditionally used in black-and-white photography, and provide the same range of tones: neutral, "warm" or "cold". When switching between full-color and monochrome ink sets, it is necessary to flush out the old ink from the print head with a cleaning cartridge. Special software or at least a modified device driver are usually required, to deal with the different color mapping.
Some types of industrial inkjet printers are now capable of printing at very high speeds, in wide formats, or for a variety of industrial applications ranging from signage, textiles, optical media, ceramics and 3-D printing into biomedical applications and conductive circuitry. Leading companies and innovators in hardware include HP, Epson, Canon, Konica Minolta, FujiFilm, EFi, Durst, Brother, Roland, Mimaki, Mutoh and many others worldwide.
Disadvantages
Many "intelligent" ink cartridges contain a microchip that communicates the estimated ink level to the printer; this may cause the printer to display an error message, or incorrectly inform the user that the ink cartridge is empty. In some cases, these messages can be ignored, but some inkjet printers will refuse to print with a cartridge that declares itself empty, to prevent consumers from refilling cartridges. For example, Epson embeds a chip which prevents printing when the chip claims the cartridge is empty, although a researcher who over-rode the system found that in one case he could print up to 38% more good quality pages, even though the chip stated that the cartridge was empty. Third-party ink suppliers sell ink cartridges at significant discounts (at least 10–30% off OEM cartridge prices, sometimes up to 95%, typically averaging around 50%), and also bulk ink and cartridge self-refill kits at even lower prices. Many vendors' "intelligent" ink cartridges have been reverse-engineered. It is now possible to buy inexpensive devices to reliably reset such cartridges to report themselves as full, so that they may be refilled many times.
The very narrow inkjet nozzles are prone to clogging. The ink consumed cleaning them—either during cleaning invoked by the user, or in many cases, performed automatically by the printer on a routine schedule—can account for a significant proportion of the ink used in the machine. Inkjet printing head nozzles can be cleaned using specialized solvents; or by soaking in warm distilled water for short periods of time, for water-soluble inks.
The high cost of OEM ink cartridges and the intentional obstacles to refilling them have been addressed by the growth of third-party ink suppliers. Many printer manufacturers discourage customers from using third-party inks, stating that they can damage the print heads due to not being the same formulation as the OEM inks, cause leaks, and produce inferior-quality output (e.g. of incorrect color gamut). Consumer Reports has noted that some third-party cartridges may contain less ink than OEM cartridges, and thus yield no cost savings, while Wilhelm Imaging Research claims that with third-party inks the lifetime of prints may be considerably reduced. However, an April 2007 review showed that, in a double-blind test, reviewers generally preferred the output produced using third-party ink over OEM ink. In general, OEM inks have undergone significant system reliability testing with the cartridge and print-head materials, whereas R&D efforts on third-party ink material compatibility are likely to be significantly less. Some inkjet manufacturers have tried to prevent cartridges being refilled using various schemes including fitting chips to the cartridges that log how much the cartridge has printed and prevent the operation of a refilled cartridge.
The warranty on a printer may not apply if the printer is damaged by the use of non-approved supplies. In the US the Magnuson–Moss Warranty Act is a federal law which states that warrantors cannot require that only brand name parts and supplies be used with their products, as some printer manufacturers imply. However, this would not apply if non-approved items cause damage. In the UK, a printer manufacturer cannot lawfully impose such conditions as part of its warranty (Regina Vs Ford Motor Company refers) although many attempt to do so illegally. As long as the product used was sold as being for the printer it was used in, then the sale of goods act applies, and anything so sold must be "of merchandisable quality and fit for purpose". Moreover, under UK law, it is the retailer and not the manufacturer that is legally liable, for 2 years on electrically operated items specifically, and as such the retailer is where one would seek redress.
Durability
Inkjet documents can have poor to excellent archival durability, depending on the quality of the inks and paper used. If low-quality paper is used, it can yellow and degrade due to residual acid in the untreated pulp; in the worst case, old prints can literally crumble into small particles when handled. High-quality inkjet prints on acid-free paper can last as long as typewritten or handwritten documents on the same paper.
Because the ink used in many low-cost consumer inkjets is water-soluble, care must be taken with inkjet-printed documents to avoid even the smallest drop of moisture, which can cause severe "blurring" or "running". In extreme cases, even sweaty fingertips during hot humid weather could cause low-quality inks to smear. Similarly, water-based highlighter markers can blur inkjet-printed documents and discolor the highlighter's tip. The lifetime of inkjet prints produced using aqueous inks is generally shorter (although UV-resistant inks are available) than those produced with solvent-based inkjets; however, so-called "archival inks" have been produced for use in aqueous-based machines which offer extended life.
In addition to smearing, gradual fading of many inks can be a problem over time. Print lifetime is highly dependent on the quality and formulation of the ink. The earliest inkjet printers, intended for home and small office applications, used dye-based inks. Even the best dye-based inks are not as durable as pigment-based inks, which are now available for many inkjet printers. Many inkjet printers now utilize pigment based inks which are highly water resistant: at least the black ink is often pigment-based. Resin or silicone protected photopaper is widely available at low cost, introducing complete water and mechanical rub resistance for dye and pigment inks. The photopaper itself must be designed for pigment or for dye inks, as pigment particles are too large to be able to penetrate through dye-only photopaper protection layer.
The highest-quality inkjet prints are often called "giclée" prints, to distinguish them from less-durable and lower-cost prints. However, the use of the term is no guarantee of quality, and the inks and paper used must be carefully investigated before an archivist can rely on their long-term durability.
To increase the durability of inkjet printer prints, more attention is needed for the inkjet ink cartridge. One way to treat ink cartridges on an inkjet printer is to maintain the temperature of the printer itself. Excessive variation in space temperature is very bad for printer ink cartridges. The user should prevent the printer becoming too hot or too chilly as the cartridges can dry up. For lasting printer efficiency, the user should ensure the area has a regular and steady temperature level.
Operating cost tradeoffs
Inkjets use solvent-based inks which have much shorter expiration dates compared to laser toner, which has an indefinite shelf life. Inkjet printers tend to clog if not used regularly, whereas laser printers are much more tolerant of intermittent use. Inkjet printers require periodical head cleaning, which consumes a considerable amount of ink, and will drive printing costs higher especially if the printer is unused for long periods.
If an inkjet head becomes clogged, third-party ink solvents/head cleaners and replacement heads are available in some cases. The cost of such items may be less expensive compared to a transfer unit for a laser printer, but the laser printer unit has a much longer lifetime between required maintenance. Many inkjet printer models now have permanently installed heads, which cannot be economically replaced if they become irreversibly clogged, resulting in scrapping of the entire printer. On the other hand, inkjet printer designs which use a disposable printhead usually cost significantly more per page than printers using permanent heads. By contrast, laser printers do not have printheads to clog or replace frequently, and usually can produce many more pages between maintenance intervals.
Inkjet printers have traditionally produced better quality output than color laser printers when printing photographic material. Both technologies have improved dramatically over time, although the best quality giclee prints favored by artists use what is essentially a high-quality specialized type of inkjet printer.
Business model
A common business model for inkjet printers involves selling the actual printer at or below production cost, while dramatically marking up the price of the (proprietary) ink cartridges (a profit model called "Freebie marketing"). Most current inkjet printers attempt to enforce this product tying by anticompetitive measures such as microchips in the cartridges to hinder the use of third-party or refilled ink cartridges. The microchips monitor usage and report the ink remaining to the printer. Some manufacturers also impose "expiration dates". When the chip reports that the cartridge is empty (or out of date) the printer stops printing. Even if the cartridge is refilled, the microchip will indicate to the printer that the cartridge is depleted. For many models (especially from Canon), the 'empty' status can be overridden by entering a 'service code' (or sometimes simply by pressing the 'start' button again). For some printers, special circuit "flashers" are available that reset the quantity of remaining ink to the maximum.
Some manufacturers, most notably Epson and Hewlett Packard, have been accused of indicating that a cartridge is depleted while a substantial amount of ink remains. A 2007 study found that most printers waste a significant quantity of ink when they declare a cartridge to be empty. Single-ink cartridges were found to have on average 20% of their ink remaining, though actual figures range from 9% to 64% of the cartridge's total ink capacity, depending on the brand and model of printer. This problem is further compounded with the use of one-piece multi-ink cartridges, which are declared empty as soon as one color runs low. Of great annoyance to many users are those printers that will refuse to print documents requiring only black ink, just because one or more of the color ink cartridges is depleted.
In recent years, many consumers have begun to challenge the business practices of printer manufacturers, such as charging up to US$8,000 per gallon (US$2,100 per liter) for printer ink. Alternatives for consumers are cheaper copies of cartridges, produced by third parties, and the refilling of cartridges, using refill kits. Due to the large differences in price caused by OEM markups, there are many companies selling third-party ink cartridges. Most printer manufacturers discourage refilling disposable cartridges or using aftermarket copy cartridges, and say that use of incorrect inks may cause poor image quality due to differences in viscosity, which can affect the amount of ink ejected in a drop, and color consistency, and can damage the printhead. Nonetheless, the use of alternative cartridges and inks has been gaining in popularity, threatening the business model of printer manufacturers. Printer companies such as HP, Lexmark, and Epson have used patents and the DMCA to launch lawsuits against third-party vendors. An anti-trust class-action lawsuit was launched in the US against HP and office supply chain Staples, alleging that HP paid Staples $100 million to keep inexpensive third-party ink cartridges off the shelves.
In Lexmark Int'l v. Static Control Components, the United States Court of Appeals for the Sixth Circuit ruled that circumvention of this technique does not violate the Digital Millennium Copyright Act. The European Commission also ruled this practice anticompetitive: it will disappear in newer models sold in the European Union. While the DMCA case dealt with copyright protection, companies also rely on patent protection to prevent copying and refilling of cartridges. For example, if a company devises all of the ways in which their microchips can be manipulated and cartridges can be refilled and patents these methods, they can prevent anyone else from refilling their cartridges. Patents protecting the structure of their cartridges prevent the sale of cheaper copies of the cartridges. For some printer models (notably those from Canon) the manufacturer's own microchip can be removed and fitted to a compatible cartridge thereby avoiding the need to replicate the microchip (and risk prosecution). Other manufacturers embed their microchips deep within the cartridge in an effort to prevent this approach.
In 2007 Eastman Kodak entered the inkjet market with its own line of All-In-One printers based on a marketing model that differed from the prevailing practice of selling the printer at a loss while making large profits on replacement ink cartridges. Kodak claimed that consumers could save up to 50 percent on printing by using its lower cost cartridges filled with the company's proprietary pigmented colorants while avoiding the potential problems associated with off-brand inks. This strategy proved unsuccessful and Kodak exited the consumer inkjet printer business in 2012.
A more recent development is the supertank printer, which uses an integrated continuous ink system. Supertank printers are defined by their large, permanently-installed ink tanks which are filled from ink bottles. The printer itself is typically sold at a substantial premium, but ink bottles are inexpensive and contain enough ink to print thousands of pages. Supertank printers generally ship with full bottles of ink in the box, allowing up to two years of printing before the tanks needs to be refilled. Epson pioneered this technology by launching the EcoTank range, first in Indonesia in 2010, with a North American launch in 2015. The supertank concept proved commercially successful, and other printer manufacturers have since launched their own lines of supertank printers, including Canon and HP (which also extended the refillable-tank concept to laser printers).
Printer types
Professional models
In addition to the widely used small inkjet printers for home and office, there are professional inkjet printers, some for "page-width" format printing and many for wide format printing. Page-width format means that the print width ranges from about . "Wide format" means print width ranging from 24" up to 15' (about 60 cm to 5 m). The most common application of page-width printers is in printing high-volume business communications that do not need high-quality layout and color. Particularly with the addition of variable data technologies, the page-width printers are important in billing, tagging, and individualized catalogs and newspapers. The application of most wide format printers is in printing advertising graphics; a lower-volume application is printing of design documents by architects or engineers. But nowadays, there are inkjet printers for digital textile printing up to 64" wide with good high definition image of 1440×720 dpi.
Another specialty application for inkjets is producing prepress color proofs for printing jobs created digitally. Such printers are designed to give accurate color rendition of how the final image will look (a "proof") when the job is finally produced on a large volume press such as a four-colour offset lithography press. An example is an Iris printer, whose output is what the French term giclée was coined for.
The largest-volume supplier is Hewlett-Packard, which supplies over 90 percent of the market for printers for printing technical drawings. The major products in their Designjet series are the Designjet 500/800, the Designjet T Printer series (including the T1100 and T610), the Designjet 1050 and the Designjet 4000/4500. They also have the HP Designjet 5500, a six-color printer that is used especially for printing graphics as well as the new Designjet Z6100 which sits at the top of the HP Designjet range and features an eight colour pigment ink system.
Epson, Kodak, and Canon also manufacture wide-format printers, sold in much smaller numbers than standard printers. Epson has a group of three Japanese companies around it that predominantly use Epson piezo printheads and inks: Mimaki, Roland, and Mutoh.
Scitex Digital Printing developed high-speed, variable-data, inkjet printers for production printing, but sold its profitable assets associated with the technology to Kodak in 2005 who now market the printers as Kodak Versamark VJ1000, VT3000, and VX5000 printing systems. These roll-fed printers can print at up to 1000 feet per minute.
Professional high-volume inkjet printers are made by a range of companies. These printers can range in price from US$35,000 to $2 million. Carriage widths on these units can range from 54" to 192" (about 1.4 to 5 m), and ink technologies have tended toward solvent, eco-solvent, and UV-curing with a more recent focus toward water-based (aqueous) ink sets. Major applications where these printers are used are for outdoor settings for billboards, truck sides and truck curtains, building graphics and banners, while indoor displays include point-of-sales displays, backlit displays, exhibition graphics, and museum graphics.
The major suppliers for professional high-volume, wide- and grand-format printers include: EFI, LexJet, Grapo, Inca, Durst, Océ, NUR (now part of Hewlett-Packard), Lüscher, VUTEk, Scitex Vision (now part of Hewlett-Packard), Mutoh, Mimaki, Roland DG, Seiko I Infotech, IQDEMY, Leggett and Platt, Agfa, Raster Printers, DGI and MacDermid ColorSpan (now part of Hewlett-Packard), swissqprint, SPGPrints (formerly Stork Prints), MS Printing Systems and Digital Media Warehouse.
SOHO multifunction inkjet photo printers
SOHO multifunction inkjet printers for photo printing use up to 6 different inks:
Canon: cyan, yellow, magenta, black, pigment black, gray. 1 pl thermal.
Epson: cyan, yellow, magenta, light cyan, light magenta, black. 1.5 pl piezo variable. Also with A3 paper printing, or FAX and duplex ADF.
Professional inkjet photo printers
Inkjet printers for professional photo printing use up to twelve different inks:
Canon: photo magenta, photo cyan, yellow, magenta, cyan, red, photo black, matte black, grey, plus either blue, photo gray, and one chroma optimiser for black density and uniform glossiness, or light gray, dark gray and one chroma optimiser, or green, blue, and photo gray. 4 pl thermal.
Epson (10 colors from 12): vivid magenta, yellow, cyan, orange, green, vivid light magenta, light cyan, light black, matte black or photo black, plus an irreversible choice of either light light gray or violet (V not for photo). 3.5 pl piezo variable.
HP: magenta, yellow, red, green, blue, light magenta, light cyan, gray, light gray, matte black, photo black, and one gloss enhancer. 4 pl thermal.
They can print an image of 36 megapixels on A3 borderless photo paper with 444 ppi.
Compact photo printers
A compact photo printer is a stand-alone inkjet printer designed to produce 4×6 or 2×3 inch prints from digital cameras. It works without the use of a computer. It is also known as a portable photo printer or a snapshot printer. Compact photo printers came on the market shortly after the popularity of home photo printing took off in the early 2000s. They were designed as an alternative to developing photos or printing them on a standard inkjet photo printer.
The majority of compact photo printers can only print 4" x 6" pictures. Given this limitation, they are not meant to replace standard inkjets. Many manufacturers advertise the cost per page of photos printed on their machines; this theoretically convinces people that they can print their own pictures just as cheaply as retail stores or through online printing services. Most compact photo printers share a similar design. They are small units, usually with large LCDs in order to allow people to browse and edit their photos, as can be done on a computer. The editing options are usually somewhat advanced, allowing the user to crop photos, remove red eye, adjust color settings as well as other functions. Compact photo printers typically feature a large number of connection options, including USB and most memory card formats.
Compact photo printers are currently manufactured by most of the leading printer manufacturers such as Epson, Canon, HP, Lexmark and Kodak. While they have increased in popularity in recent years, they still make up a relatively small share of the inkjet printer market. LG's Pocket Photo uses Zink thermal paper which has chemistries embedded on each inkless photo paper and the image will appear with the heat.
Other uses
U.S. Patent 6,319,530 describes a "Method of photocopying an image onto an edible web for decorating iced baked goods". In other words, this invention enables one to inkjet print a food-grade color photograph on a birthday cake's surface. Many bakeries now carry these types of decorations, which are printable using edible inks and dedicated inkjet printers. Edible ink printing can be done using normal home use inkjet printers like Canon Bubble Jet printers with edible ink cartridges installed, and using rice paper or frosting sheets.
Inkjet printers and similar technologies are used in the production of many microscopic items. See Microelectromechanical systems.
Inkjet printers are used to form conductive traces for circuits, and color filters in LCD and plasma displays.
Inkjet printers, especially models produced by Dimatix (now part of Fujifilm), Xennia Technology and Pixdro, are in fairly common use in many labs around the world for developing alternative deposition methods that reduce consumption of expensive, rare, or problematic materials. These printers have been used in the printing of polymer, macromolecular, quantum dot, metallic nanoparticles, and carbon nanotubes. The applications of such printing methods include organic thin-film transistors, organic light emitting diodes, organic solar cells, and sensors.
Inkjet technology is used in the emerging field of bioprinting. They are also used for the production of OLED displays.
See also
Edgeline printing
Inkjet solar cell
Inkjet technology
Inkjet transfer
Intelligent Interweaving technology
ISO Standards for colour ink jet printers
Laser printing
Memjet
Microfluidics
Society for Imaging Science and Technology
Tonejet
UV pinning
Water-jet printer
References
Further reading
External links
Inkjet printing, information video from the University of Sheffield
Computer-related introductions in 1976
Office equipment |
null | null | Princess Leia | eng_Latn | Princess Leia Organa is a fictional character in the Star Wars franchise, portrayed in films by Carrie Fisher. Introduced in the original Star Wars film in 1977, Leia is princess of the planet Alderaan, a member of the Imperial Senate and an agent of the Rebel Alliance. She thwarts the sinister Sith lord Darth Vader and helps bring about the destruction of the Empire's cataclysmic superweapon, the Death Star. In The Empire Strikes Back (1980), Leia commands a Rebel base and evades Vader as she falls in love with the smuggler Han Solo. In Return of the Jedi (1983), Leia helps in the operation to rescue Han from the crime lord Jabba the Hutt and is revealed to be Vader's daughter and the twin sister of Luke Skywalker.
The prequel film Revenge of the Sith (2005) establishes that the twins' mother is Senator (and former queen) Padmé Amidala of Naboo, who dies after childbirth, while their father is none other than former Jedi Knight Anakin Skywalker, who would become Darth Vader. Leia is adopted by Senator Bail and Queen Breha Organa of Alderaan. In the sequel trilogy, Leia is the founder and General of the Resistance against the First Order. She and Han have a son named Ben Solo, who adopted the name Kylo Ren after turning to the dark side of the Force and became the lead enforcer for the First Order. In The Rise of Skywalker (2019), it is revealed that Leia was briefly trained as a Jedi by her brother sometime after Return of the Jedi. In the film, Leia is the mentor of Rey, the last remaining Jedi. Leia dies towards the end of the film but returns in the epilogue scene as a Force ghost alongside Luke.
One of the more popular Star Wars characters, Leia has been called a 1980s icon, a feminist hero and model for other adventure heroines. She has appeared in many derivative works and merchandising, including the now-noncanonical Star Wars Expanded Universe, and has been referenced or parodied in several TV shows and films. Her 'cinnamon bun' hairstyle from Star Wars (1977) and metal bikini from Return of the Jedi have become cultural icons. Fisher was nominated for the Saturn Award for Best Actress twice for A New Hope and Return of the Jedi. Fisher later received two Saturn Award nominations for Best Supporting Actress, the first for The Force Awakens and the second being a posthumous nomination for The Last Jedi.
Creation and casting
Leia was created by Star Wars creator George Lucas, who in 1999 explained his early development of the main characters:
In the rough draft of Star Wars, Leia is the spoiled teenage daughter of King Kayos and Queen Breha of Aquilae, with two brothers, Biggs and Windy; Biggs returned to the fourth draft as a childhood friend of Luke. Leia was at one point "the daughter of Owen Lars and his wife Beru ... Luke's cousin–together they visit the grave of his mother, who perished with his father on a planet destroyed by the Death Star." A later story synopsis establishes Leia as "Leia Antilles", the daughter of Bail Antilles from the peaceful world of Organa Major. In the fourth draft it was established that "Leia Organa" came instead from Alderaan.
Fisher was 19 when she was cast as Princess Leia, with actresses including Amy Irving, Cindy Williams and Jodie Foster also up for the role. In 2014, InkTank reported that the extended list of "more than two dozen actresses" who had auditioned for Leia included Glenn Close, Farrah Fawcett, Jessica Lange, Sissy Spacek, Sigourney Weaver, Cybill Shepherd, Jane Seymour, Anjelica Huston, Kim Basinger, Kathleen Turner, Geena Davis, Meryl Streep, and Terri Nunn. Asked about Streep in 2015, Fisher said, "I've never heard that one. But Jodie Foster was up for it... that one I knew the most. Amy Irving and Jodie. And I got it."
On Leia's 'cinnamon bun' hairstyle, George Lucas stated that "In the 1977 film, I was working very hard to create something different that wasn't fashion, so I went with a kind of Southwestern Pancho Villa woman revolutionary look. ... The buns are basically from turn-of-the-century Mexico." Lucas later noted that Leia's mother, Padmé Amidala, wears a similar-looking Hopi hairstyle in Revenge of the Sith. A photo depicting famous Mexican revolutionary Clara de la Rocha wearing a similar hairstyle exists in the Lucas Museum of Narrative Art archives. In 2015, Mental Floss cited research suggesting that while female Mexican revolutionaries did not have such elaborate hairstyles, young marriageable Hopi women did, and their "squash blossom whorls" superficially resemble Leia's hair buns. Leia's hairstyle may have also been inspired by that of Queen Fria in the 1939 Flash Gordon comic strip "The Ice Kingdom of Mongo", and scientist Barnes Wallis' wife Molly (played by Ursula Jeans) in the 1955 war film The Dam Busters. Comparisons have also been made to the 4th-century-BC Iberian sculpture Lady of Elche, as well as the 1920s "earphones" hairstyle.
Composer John Williams created a musical leitmotif for Leia which recurs throughout the Star Wars saga. "Princess Leia's Theme" was recorded as a concert suite (4:18 length) for the score of the 1977 film.
The second draft of the Return of the Jedi screenplay contained dialogue in which Obi-Wan tells Luke he has a twin sister. She and their mother were "sent to the protection of friends in a distant system. The mother died shortly thereafter, and Luke's sister was adopted by Ben's friends, the governor of Alderaan and his wife." Fisher explained in 1983: "Leia's real father left her mother when she was pregnant, so her mother married this King Organa. I was adopted and grew up set apart from other people because I was a princess."
Character
Anthony Breznican of Entertainment Weekly describes Leia as a "diplomat, spy, warrior, undercover agent". Mark Edlitz calls her "a smart, feisty, brave diplomat and warrior" in The Huffington Post. Fisher told Rolling Stone in 1983:
She said in 2014:
Appearances
Skywalker saga
Original trilogy
Star Wars (A New Hope)
Introduced in the original 1977 film Star Wars, Princess Leia Organa of Alderaan is a member of the Imperial Senate. She is captured by Darth Vader (portrayed by David Prowse, voiced by James Earl Jones) on board the ship Tantive IV, where she is acting as a spy for the Rebel Alliance. Leia has secretly hidden the blueprints for the Death Star, the Empire's moon-sized battle station, inside the astromech droid R2-D2 (Kenny Baker) and has sent it to find one of the last remaining Jedi, Obi-Wan Kenobi (Alec Guinness), on the nearby planet of Tatooine. Vader arrests Leia and has her tortured, but she resists revealing anything. Death Star commander Grand Moff Tarkin (Peter Cushing) threatens to destroy her home planet Alderaan with the Death Star unless she reveals the location of the hidden Rebel base. She provides the location of an old, abandoned base on the planet Dantooine, but Tarkin orders Alderaan to be destroyed anyway. Leia is rescued by Luke Skywalker (Mark Hamill) and Han Solo (Harrison Ford), and they escape aboard Han's ship, the Millennium Falcon. After analyzing the Death Star's schematics, the Rebels find a tiny weakness in the battle station, which Luke uses to destroy it in his X-wing. In the aftermath of the victory, Leia bestows medals on the heroes at the hidden Rebel base on Yavin 4.
Alyssa Rosenberg of The Washington Post writes of Leia in the film:
Rosenberg also notes that, though Han is almost immediately attracted to Leia, they conflict because she insists on asserting command and he automatically resists, even as she proves herself to be worthy of it. And despite her initial disdain for the smuggler, whom she sees as selfish, Leia later acknowledges "I knew there was more to you than money" when Han comes through for the Rebellion.
Fisher told Rolling Stone in 1980 that in the original script, when Luke and Han come to rescue a captured Leia, "I was hanging upside down with yellow eyes, like in The Exorcist ... Some form of radar torture was done to me and I was in a beam, bruised and beaten up, suspended in midair. The reason it was cut from the film was because I was unconscious and the Wookiee would have had to carry me for, like, the next fifteen minutes."
The Empire Strikes Back
In The Empire Strikes Back (1980), Leia is at the Rebel base on Hoth. She leads its evacuation during an Imperial attack, and then flees with Han, Chewbacca (Peter Mayhew) and C-3PO (Anthony Daniels) on the Millennium Falcon. They dodge pursuing Imperial TIE fighters by flying into an asteroid field when the Falcon hyperdrive breaks down, with Leia piloting the ship at one point. Romance blossoms between Leia and Han during their flight from the Empire; while hiding in the stomach of a space slug, they finally share a kiss. With his ship needing repairs, Han seeks out his old friend Lando Calrissian (Billy Dee Williams) in Cloud City, the floating city over Bespin. Though he welcomes them graciously, Lando soon turns them over to a newly arrived Darth Vader, who hopes to use them as bait to capture Luke. Leia confesses her love for Han as he is frozen in carbonite and then handed over to bounty hunter Boba Fett (portrayed by Jeremy Bulloch, voiced by Jason Wingreen), who is charged with bringing him to the crime lord Jabba the Hutt. Lando helps Leia, Chewbacca and the two droids escape. Leia senses that Luke is in trouble, and goes back to save him after he is nearly killed during a lightsaber duel and confrontation with Vader.
Rosenberg notes:
Return of the Jedi
In Return of the Jedi (1983), Leia infiltrates Jabba the Hutt's palace on Tatooine disguised as the Ubese bounty hunter Boushh and brings a captive Chewbacca with her as part of the ruse. She frees Han from the carbonite, but they are recaptured by Jabba. Leia is now chained to Jabba as his slave, outfitted in a metal bikini. After Luke arrives and kills Jabba's Rancor, Jabba sentences Luke, Han and Chewbacca to be fed to the Sarlacc. Lando (disguised as a guard) helps them overpower their captors, and Leia seizes the moment to strangle Jabba with her chain. Luke and Leia swing to safety, blowing up Jabba's barge behind them. Later, the heroes go to the forest moon of Endor to prepare for a battle with the Empire. There, Luke reveals to Leia that she is his twin sister and that Vader is their father. Leia joins Han in leading the Rebels in battle with Imperial troops as the Rebel fleet attacks the second Death Star. Leia is slightly injured, but the Rebels, with help from the Ewoks, ultimately defeat the forces of the Empire, with the Death Star destroyed once more.
Fisher told Rolling Stone in 1983, "In Return of the Jedi, [Leia] gets to be more feminine, more supportive, more affectionate. But let's not forget that these movies are basically boys' fantasies. So the other way they made her more female in this one was to have her take off her clothes." Rosenberg writes of Han and Leia:
In the film, Leia says that she has vague memories of her real mother, who she describes as "kind... but sad". Bouzereau quoted Lucas in 1997:
Prequel trilogy
Revenge of the Sith
In the prequel film Revenge of the Sith (2005), Padmé Amidala (Natalie Portman) is pregnant with Anakin Skywalker's (Hayden Christensen) twins near the end of the Clone Wars. After Anakin turns to the dark side of the Force and would become Darth Vader, would Padmé gives birth to Luke and Leia on Polis Massa and then dies. Leia is adopted by Senator Bail Organa (Jimmy Smits) and his wife Queen Breha (Rebecca Jackson Mendoza) of Alderaan.
Film critic Peter Travers of Rolling Stone wrote, "As we watch Anakin nearly melt in the lava, only to be put together, Frankenstein style, in a lab while Lucas intercuts scenes of Padme giving birth to the twins Luke and Leia, a link to genuine feeling is established at last."
Sequel trilogy
The Force Awakens
In March 2013, Fisher confirmed that she would reprise her role as Leia in Star Wars: The Force Awakens. Set 30 years after Return of the Jedi, The Force Awakens reintroduces a Leia who is "a little more battle weary, a little more broken hearted". In November 2015, director J. J. Abrams said of Leia, "She's referred to as General but ... there's a moment in the movie where a character sort of slips and calls her 'Princess.'" Commenting on the story he added, "The stakes are pretty high in the story for her, so there's not much goofing around where Leia's concerned." Asked how Leia is handling things in the film, Fisher said, "Not easily ... [she is] solitary. Under a lot of pressure. Committed as ever to her cause, but I would imagine feeling somewhat defeated, tired, and pissed."
In the film, Leia is the leader of the Resistance, a group formed by the New Republic to fight a proxy war with the First Order (a group formed by the remnants of the Empire) and trying to find Luke, who disappeared years earlier. When her forces foil a First Order attack searching for the Resistance astromech droid BB-8, Leia is reunited with Han, who has helped the renegade stormtrooper Finn (John Boyega) and orphaned scavenger Rey (Daisy Ridley) bring the droid this far. Han and Leia are still married but have been estranged for a couple of years; after their son, Ben Solo (Adam Driver), fell to the dark side and became the First Order high-ranking member Kylo Ren. Leia believes Ben can still be brought back to the light side. Han volunteers for a mission to infiltrate the First Order's Starkiller Base to disable its defensive shields. Confronting Ren, Han tries to convince his son to leave the First Order to honor Leia's request of bringing him home, but Ren instead impales his father on his lightsaber. Leia senses Han's death through the Force, and later shares a moment of grief with Rey, who had thought of Han as a mentor and father figure. Leia sees Rey, Chewbacca, and R2-D2 off when they depart to locate Luke, saying, "may the Force be with you".
Asked why Leia is not shown to be a Jedi in The Force Awakens (as she is in the Star Wars Expanded Universe works), Abrams told IGN, "It was a great question, and one that we talked about quite a bit, even with Carrie [Fisher]. Why did she not take advantage of this natural Force strength that this character had. And one of the answers was that it was simply a choice that she made, that her decision to run the Rebellion, and ultimately this Resistance, and consider herself a General, as opposed to a Jedi. It was simply a choice that she took". He also added, "I would like to think that there really isn't much of a ticking clock, and it's never too late ... clearly we've seen, and we do again, that she still is Force strong. And it's something that I think is an intrinsic piece of her character.
Fisher was nominated for a 2016 Saturn Award for Best Supporting Actress for her portrayal.
The Last Jedi
In December 2015, producer Kathleen Kennedy confirmed that Fisher would reprise the role of Leia in the next installment, then known as Star Wars: Episode VIII, but later named Star Wars: The Last Jedi. Fisher died on December 27, 2016, after going into cardiac arrest. It was confirmed that she had completed filming her role in the film shortly before her death.
In the film, Leia is among those on the bridge of her flagship, the MC85 Star Cruiser Raddus, who are expelled into space when the ship is attacked by the First Order. Leia uses the Force to pull herself back to the ship. After recovering, she shoots and stuns Resistance pilot Poe Dameron (Oscar Isaac), who has mutinied against her successor, Vice Admiral Holdo (Laura Dern). From his solitude on Ahch-To, Luke projects himself through the Force to the Resistance stronghold on Crait and reunites with Leia, apologizing for what happened to Ben. Leia replies that she knows her son is gone, but Luke reassures her that "no one's ever really gone". While Luke distracts Kylo and his attacking First Order troops, Leia is among the remaining Resistance forces who escape from Crait in the Millennium Falcon.
The Rise of Skywalker
Following Fisher's death, Variety reported that she was slated to appear in Star Wars: The Rise of Skywalker and speculated that Lucasfilm would need to find a way to address her death and what would become of her character. Filming began on August 1, 2018. Lucasfilm announced in January 2017 that they had "no plans to digitally recreate Carrie Fisher's performance as Princess or General Leia Organa". In April 2017, Fisher's family granted Disney and Lucasfilm the rights to use recent footage of the late actress in The Rise of Skywalker. However, Kennedy subsequently said that Fisher would not appear in the film. In July 2018, it was announced that Fisher would appear as Leia in The Rise of Skywalker, using unseen footage from The Force Awakens. Initially, there had also been plans to use additional unseen footage of Leia from The Last Jedi, but it was ultimately not included in the ninth film. Fisher's daughter Billie Lourd, who portrayed Lieutenant Connix in all three sequel trilogy films, also stepped in as Leia for a brief flashback in the film in which her face was digitally replaced by Fisher's likeness, using imagery from Return of the Jedi.
In the film, Leia continues to lead the Resistance while offering guidance and support to Rey as she continues her training to become a Jedi. A flashback reveals that Leia had abandoned her own training after she had a vision foretelling her son's death if she finished. While Rey and Ren duel on Kef Bir, a dying Leia uses all of her remaining strength to reach out to her son. Distracted after sensing his mother's death, Ren is stabbed by Rey with his own lightsaber. Leia passes away, making Rey experience overwhelming guilt. Rey heals Ren's wound using the Force. After the Battle of Exegol, a redeemed Ben Solo sacrifices the remainder of his life force to resurrect a dead Rey, and he vanishes at the same time as Leia becomes one with the Force. Rey then returns to the Lars homestead on Tatooine and buries the lightsabers that had belonged to Leia and her father, Anakin Skywalker as the spirits of Luke and Leia look on with pride, with Rey honoring them by adopting the surname "Skywalker".
Anthology films
Rogue One: A Star Wars Story
Leia makes a brief appearance in the final scene of the 2016 film Rogue One, receiving the plans for the Death Star as a lead-up to the beginning of A New Hope. Since this movie takes place prior to the original Star Wars trilogy, a very young Leia was required. To achieve that effect, a computer-generated image of a young Carrie Fisher was superimposed over Norwegian actress Ingvild Deila's face; archival audio of Fisher saying "Hope" was used to voice the character.
Television
Leia appears briefly in 1978's Star Wars Holiday Special television film as a leader and administrator of the new Rebel Alliance base. She and C-3PO contact Chewbacca's wife Mallatobuck for assistance in finding Chewbacca and Han. Leia also appears in the cartoon segment at a different Rebel Base, located in an asteroid field, and at the Life Day ceremony at the end of the film. Fisher also appeared in and hosted the November 18, 1978, episode of Saturday Night Live that aired one day after the holiday special.
Star Wars Rebels
A teenage version of Princess Leia, voiced by Julie Dolan, appears in a 2016 episode of the animated series Star Wars Rebels, which is set between Revenge of the Sith and A New Hope. In the episode, Leia is sent on a secret mission to assist the titular rebels. Executive producer Dave Filoni said of the appearance:
Star Wars Forces of Destiny
Leia appears in the animated series Star Wars Forces of Destiny, voiced by Shelby Young.
Star Wars Resistance
Leia appears in the animated series Star Wars Resistance, voiced by Rachel Butera and Carolyn Hennesy.
Novels
Leia makes her first literary appearance in Star Wars: From the Adventures of Luke Skywalker, the novelization of the original 1977 film Star Wars, which was released six months before the film in November 1976. Credited to Lucas but ghostwritten by Alan Dean Foster, the novel was based on Lucas' screenplay. Leia later appeared in the novelizations The Empire Strikes Back (1980) by Donald F. Glut and Return of the Jedi (1983) by James Kahn. She is also a point-of-view character in the 2015 novelization of The Force Awakens by Foster.
Foster's 1978 novel Splinter of the Mind's Eye was commissioned by Lucas as the basis for a potential low-budget sequel to Star Wars should the film prove unsuccessful. In the story, Luke and Leia seek a crystal on a swampy planet and eventually face Vader in combat.
Leia also appears in the Journey to Star Wars: The Force Awakens line of novels and comic books, introduced in conjunction with The Force Awakens to connect the film with previous installments. She is the lead character in the young adult novel Moving Target: A Princess Leia Adventure (2015) by Cecil Castellucci and Jason Fry, which is set between The Empire Strikes Back and Return of the Jedi, and Claudia Gray's novels Star Wars: Bloodline (2016) and Leia: Princess of Alderaan (2017). The former is set six years before The Force Awakens, while the latter features a 16-year-old Leia before the events of A New Hope.
Comics
Leia is the lead character in the five-part comic limited series Star Wars: Princess Leia (2015), taking place immediately after Episode IV: A New Hope. She is also featured prominently in the four-part comic limited series Star Wars: Shattered Empire (2015), set immediately after Return of the Jedi. Princess Leia reveals Leia to have had royal training in martial arts and explores her reaction to the destruction of Alderaan, while Shattered Empire portrays her as a skilled pilot who undertakes a dangerous mission alongside Poe Dameron's mother. In Princess Leia and Shattered Empire, Leia senses the past while on Naboo (briefly envisioning her mother and Darth Maul, respectively), and in Star Wars #12, she uses a lightsaber as a weapon for the first time, canonically.
Legends works
The original three Star Wars films have spawned a large franchise of works that include novels, comic books, and video games. Leia appears in much of this material. In April 2014 (with the sequel film The Force Awakens in production), Lucasfilm excluded the Star Wars Expanded Universe from official Star Wars canon, rebranding it as Star Wars Legends.
In this continuity, Leia continues her adventures with Han and Luke after Return of the Jedi, fighting Imperial resurgences and new threats to the galaxy. She becomes the Chief of State of the New Republic and a Jedi master, and is the mother to three children by Han: Jaina, Jacen and Anakin Solo.
Novels
The 1991 New York Times bestselling novel Heir to the Empire by Timothy Zahn began what would become a large collection of works set before, between and especially after the original films.
Post-Return of the Jedi
The bestselling Thrawn trilogy (1991–93) by Timothy Zahn begins five years after the events of Return of the Jedi. In Heir to the Empire (1991), Leia is married to Han and three months pregnant with twins. Noghri commandos repeatedly attempt to kidnap her as part of Grand Admiral Thrawn's plan to restore the Empire and crush the New Republic. In Dark Force Rising (1992), Leia realizes that Darth Vader and the Empire deceived the Noghri to secure their allegiance, and by revealing the truth she turns the alien race to the side of the New Republic. At one point, she remembers her adoptive parents on Alderaan, implying that the "mother" she discussed with Luke on Endor was Queen Breha. In The Last Command (1993), Leia gives birth to the twins Jaina and Jacen Solo on Coruscant during Thrawn's siege.
Leia, now the Chief of State of the New Republic, is a minor character in the Jedi Academy trilogy (1994) by Kevin J. Anderson, set after the Thrawn trilogy. Next in the timeline is the Callista trilogy: Children of the Jedi (1995) by Barbara Hambly, Darksaber (1995) by Anderson and Planet of Twilight (1997) by Hambly. In The Crystal Star (1994) by Vonda McIntyre, young Jacen, Jaina and their three-year-old brother Anakin are kidnapped in a plot to restore the Empire, but are rescued by Leia and Chewbacca. Leia struggles with the responsibilities of her position in The Black Fleet Crisis trilogy (1996) by Michael P. Kube-McDowell. In The New Rebellion (1996) by Kristine Kathryn Rusch, she avoids an assassination attempt and then aids in the defeat of the Dark Jedi Kueller, whom she shoots to death. The Corellian trilogy (1995) by Roger MacBride Allen finds Han and Leia swept up in a civil war while visiting his homeworld of Corellia with their children. In the two Hand of Thrawn novels by Timothy Zahn (1997's Specter of the Past and 1998's Vision of the Future), Leia tries to hold the New Republic together as Moff Disra conspires for its volatile factions to destroy each other. Leia appears periodically in the Young Jedi Knights series (1995–98) by Kevin J. Anderson and Rebecca Moesta. The 14-volume young adult fiction series covers the Jedi training of Jacen and Jaina.
In The Truce at Bakura (1993) by Kathy Tyers, set one day after the ending of Return of the Jedi, Leia establishes New Alderaan, a sanctuary for the destroyed planet's surviving inhabitants. The spirit of Anakin Skywalker appears to Leia and pleads for her forgiveness, but she angrily banishes him. The six-volume Jedi Prince series (1992–93) by Paul Davids and Hollace Davids, later contradicted by other novels, is set within a year after Return of the Jedi. In The Glove of Darth Vader (1992), the self-proclaimed son of the defeated Emperor Palpatine, Trioculus, seeks the titular glove to cement himself as the new Emperor. Entranced by Leia's beauty in The Lost City of the Jedi (1992), Trioculus vows to make her his queen. He captures her in Zorba the Hutt's Revenge (1992), but Jabba the Hutt's vengeful father, Zorba, offers to trade his own prisoner Ken—Palpatine's real grandson whom Trioculus has been seeking—for Leia, his son's killer. But Leia and Ken are rescued and Trioculus is frozen in carbonate by Zorba. Mission from Mount Yoda (1993) finds Ken's father Triclops alive and willing to join the Rebels against the Empire. Leia, now engaged to Han, is captured by Zorba in Queen of the Empire (1993). Trioculus is revived and seizes Leia before Zorba can kill her. Leia is rescued and replaced with a lookalike droid decoy, which kills Trioculus. In Prophets of the Dark Side (1993), Leia looks forward to her wedding to Han and has a vision of their two children. Matthew Stover's 2008 standalone novel Luke Skywalker and the Shadows of Mindor picks up the story soon after, as Luke, Leia and the Rebels fight the Sith Lord Shadowspawn.
In The Courtship of Princess Leia (1994) by Dave Wolverton, set immediately before the Thrawn trilogy, Leia is presented with an advantageous political marriage to Prince Isolder of the planet Hapes. A jealous Han abducts Leia and takes her to the planet Dathomir; Luke and Isolder follow, and there they all find the hidden forces of the Imperial warlord Zsinj. Defeating him, Han and Leia marry. The 2003 novels A Forest Apart and Tatooine Ghost by Troy Denning are set immediately after The Courtship of Princess Leia. The newly married Leia fears that any children she has may succumb to the dark side like her father. During an adventure on Tatooine in Tatooine Ghost, she discovers the diary of her grandmother Shmi Skywalker and meets some of young Anakin's childhood friends. When she learns of Anakin's childhood as a slave and the traumatic death of his mother, Leia learns to forgive her father.
Works set between films
In Shadows of the Empire (1996) by Steve Perry, the only Star Wars novel set between The Empire Strikes Back and Return of the Jedi, Leia is searching for Boba Fett to find a captive Han. She is bewitched by the crime lord Prince Xizor using pheromones, but Chewbacca helps her elude the seduction.
Allegiance (2007) and Choices of One (2011) by Timothy Zahn are set between Star Wars: A New Hope and The Empire Strikes Back, and feature Leia and her cohorts seeking new allies for their Rebellion against the Empire. Razor's Edge (2013) by Martha Wells and Honor Among Thieves (2014) by James S. A. Corey take place in the same time period and also chronicle the adventures of Leia and Han.
New Jedi Order
In the New Jedi Order series (1999–2003), Leia resigns as Chief of State, and on the heels of her warnings before the Senate, the alien Yuuzhan Vong invade the galaxy. They destroy system after system and defeat both the Jedi and the New Republic forces in countless battles. Chewbacca dies in Vector Prime (1999) by R.A. Salvatore, which sends Han into a deep depression that causes a rift between him and Leia. They reunite after Leia is gravely wounded at the Battle of Duro in Kathy Tyers' Balance Point (2000). She is targeted by a deadly Voxyn slayer in Troy Dennings' Star By Star (2001), and though she manages to evade death, her son Anakin is later killed during a mission to prevent more Voxyn from being cloned. The Vong are finally defeated in The Unifying Force (2003) by James Luceno.
In Denning's The Dark Nest trilogy (2005), Leia, Han, and several Jedi become involved in an escalating border dispute between the Chiss and the insidious insectoid Killiks, and Leia makes a bitter enemy in the Twi'lek warrior Alema Rar. In The Joiner King (2005), Leia asks Saba Sebatyne to train her as a Jedi Knight. R2-D2 malfunctions in The Unseen Queen (2005) and shows Luke a holoclip of his father Anakin and a pregnant woman, whom Luke learns is his and Leia's biological mother, Padmé Amidala. Anakin and Padmé are discussing a dream of Anakin's in which Padmé dies in childbirth; later, Luke and Leia watch a clip in which Padmé is talking to Obi-Wan Kenobi about Anakin. Tenel Ka, Queen Mother of the Hapes Consortium, has a daughter, Allana, secretly fathered by Jacen. In The Swarm War (2005), Luke and Leia view holoclips of their mother's death, and Leia is promoted to Jedi Knight.
Legacy of the Force
The bestselling Legacy of the Force series (2006–08) chronicles the crossover of Han and Leia's son Jacen to the dark side of the Force while the Jedi, Solos, and Skywalkers fight against his growing power. In Betrayal (2006) by Aaron Allston, Jacen turns to the dark side, believing that it is the only way to save the galaxy from the chaos brewing among the member systems of the Galactic Alliance. Jacen realizes in Bloodlines (2006) by Karen Traviss that the Sith discipline will require him to kill one of his loved ones, which he decides is an acceptable sacrifice to save the galaxy. In Troy Denning's Tempest (2006), Han and Leia thwart the assassination of Tenel Ka and Allana but become caught up in a Corellian conspiracy. They are almost killed when the Millennium Falcon is attacked by a Star Destroyer controlled by an increasingly powerful Jacen—who knows that his parents are on board. With Han injured, Leia and Lando further investigate the Corellians in Aaron Allston's Exile (2007), but Alema reappears to exact her vengeance on Leia. Sacrifice (2007) by Karen Traviss finds Leia and Han on the run, hunted by Jacen as traitors to the Galactic Alliance. He kills Luke's wife Mara Jade as his final sacrifice to become Darth Caedus, the new ruler of the Sith. In Inferno (2007) by Troy Denning, Han and Leia are faced with the reality that their son, now Joint Chief of State, is the enemy. Leia attempts unsuccessfully to manipulate Jacen in Aaron Allston's Fury (2007) so that the Jedi can both thwart him and neutralize Alema. Finally, in Invincible (2008) by Troy Denning, Jaina kills Jacen in a lightsaber duel. At Tenel Ka's request, Leia and Han adopt Allana, disguised with the name "Amelia" to protect her from any future vengeance against Cadeus or the Hapes Consortium. Multiple novels in the series made the New York Times Best Seller list.
The nine-volume Fate of the Jedi series (2009–12) by Aaron Allston, Troy Denning, and Christie Golden finds Han and Leia become caught up in the intensifying conflict between the Galactic Alliance and the Jedi. In the wake of Darth Cadeus' death, the now-peaceful Galactic Alliance harbors a growing mistrust toward the Jedi, and the situation is worsened by a Force-induced psychosis that begins afflicting individual Jedi, sending them on violent rampages.
In Millennium Falcon (2008) by James Luceno, set between Legacy of the Force and Fate of the Jedi, a mysterious device hidden inside the eponymous spacecraft sends Han, Leia and Allana on an adventure to investigate the ship's past before it came into Han's possession. Troy Denning's Crucible (2013), set after Fate of the Jedi and the last novel to date in the Star Wars Legends chronology, reunites Leia, Han and Luke with Lando as they aid him to thwart a vast criminal enterprise threatening his asteroid mineral refinery in the Chilean Rift nebula.
Comics
Leia's youth is depicted in the non-canon Star Wars Tales story, The Princess Leia Diaries. In it, she develops her disdain for the Empire, as well as a conflict with Tarkin. She discovers and decides to support the Rebellion.
Dark Empire
During the events of the comic series Dark Empire (1991–92), Palpatine has been resurrected in a young clone body and seduces Luke to the dark side of the Force as part of his plan to restore the Empire. A captive Leia, resisting Palpatine's attempts to turn her as well, escapes with an artifact he needs to secure his power, the Jedi Holocron. Luke pursues her, and Leia manages to turn him back. Brother and sister then fight Palpatine with the light side of the Force, turning his own Force-generated storm against him and destroying Palpatine and his Star Destroyer. In Dark Empire II (1994–95), Leia gives birth to a third child by Han, whom she names Anakin. Palpatine is reborn in an inferior, rapidly deteriorating clone body in Empire's End (1995), and seeks to possess the body of the infant Anakin.
Video games
Leia appears as a playable character in every Lego Star Wars video game to date, including the upcoming Lego Star Wars: The Skywalker Saga.
She is also a playable character in both Star Wars Battlefront and Star Wars Battlefront II.
Cultural impact
Princess Leia has been called a 1980s icon, a feminist hero and "an exemplary personification of female empowerment". In 2008, Leia was selected by Empire magazine as the 89th greatest film character of all time, and IGN listed her as their 8th top Star Wars hero. UGO Networks listed Leia as one of their best heroes of all time in 2010.
The character has been referenced or parodied in several TV shows and films, and celebrated in cosplay. Fisher appeared in the Leia metal bikini on the cover of the Summer 1983 issue of Rolling Stone, and a painting of Leia and other characters surrounding Lucas appeared on the cover of the May 25, 1983, issue of Time announcing Return of the Jedi. In 2013, cartoonist Jeffrey Brown published the bestselling Star Wars: Vader's Little Princess, a comic strip-style book featuring Darth Vader and a young Leia in humorous father-daughter situations. Princess Leia appears on a 2007 US postage stamp and a 2015 UK stamp.
Leia has also been used in a wide range of Star Wars merchandise, including statuettes, action figures and other toys, household items and clothing, office supplies, food products, and bubble bath and shampoo in Leia-shaped bottles with her head as the cap. In her one-woman show Wishful Drinking, Fisher called the Princess Leia Pez dispenser one of the "merchandising horrors" of the series. In a 2011 interview, Fisher said:
After the 2012 acquisition of LucasFilm by the Walt Disney Company, the Disney Store stated in May 2014 that the company had "no plans for Leia products". After public criticism, Disney told Time in June 2014 that it would be releasing several Leia products. Funko has since produced several versions of Leia (at least one for each film) in their POP! line of 4.5-inch vinyl figures in the Japanese super deformed style. Hasbro is set to release an action figure of Leia as she appears in the Star Wars Rebels animated series.
"Cinnamon buns" hairstyle
Leia's unique hairdo in 1977's A New Hope has come to be known as the "doughnut" or "cinnamon buns" hairstyle, and is iconic of the character and series. A February 1978 cover story for the British teen magazine Jackie included step-by-step instructions on how to replicate Leia's hair buns. In the 1978 short film parody Hardware Wars, Princess Anne-Droid has actual cinnamon buns on the side of her head. Miss Piggy of The Muppet Show copied the hairdo in a Star Wars-themed episode of the series in February 1980. In the 1987 Mel Brooks comedy film Spaceballs, Princess Vespa (Daphne Zuniga) appears to have the hairstyle, which is soon revealed to in fact be a large pair of headphones. In the parody film Thumb Wars, the role of Leia was filled by a character named Princess Bunhead, who has two cinnamon rolls for hair. In 2015, Fisher's daughter Billie Lourd's character in the horror-comedy TV series Scream Queens, a rich and disaffected sorority girl known as Chanel No. 3, wears earmuffs in every scene as an homage to Fisher's iconic Leia hairstyle. Lourd also has a cameo in The Force Awakens (2015) in which she wears Leia's distinctive hair buns.
Feminist analysis
Leia has been the subject of feminist analysis. Mark Edlitz wrote for The Huffington Post in 2010 that "Leia is an exemplary personification of female empowerment." David Bushman, television curator at the Paley Center for Media, said in 2012, "From the male perspective ... Princess Leia was a very creditable character for her time—not perfect, but certainly defiant, assertive, and strong." Alyssa Rosenberg of The Washington Post wrote in 2015, "Leia wasn't just the first great heroine of science fiction and fantasy to capture my imagination. She was one of the first characters I encountered whose power came from her political conviction and acumen." In her 2007 article "Feminism and the Force: Empowerment and Disillusionment in a Galaxy Far, Far Away", Diana Dominguez cited Leia as a welcome change from the previous portrayals of women in film and TV. She wrote:
Rosenberg writes that, though at first Luke is an apolitical innocent in search of adventure and Han is a detached opportunist in search of money, both are "influenced by Leia's passion [and] take their places as full participants in the Rebellion". She notes, "Everyone else eventually comes around to Leia's view of the world." Leia herself, singularly dedicated to her political movement against the Empire, "finds a partner in Han, acknowledging that personal happiness can help her sustain her commitment to building a better galactic order". Rosenberg cites "Leia's willingness to see the best in him, and Han's desire to live up to her belief in him" as a foundation of their relationship, also pointing out his attempts to make her recognize that she has needs like anyone else and should acknowledge that she needs him.
In their 2012 essay "Lightsabers, Political Arenas, and Marriages", Ray Merlock and Kathy Merlock Jackson cite Leia as the successor of earlier science fiction heroines Wilma Deering of Buck Rogers and Dale Arden of Flash Gordon, and the embodiment of "a new stage in the ongoing presentation of the fairy-tale princess in jeopardy". Writing that "after Leia, no longer would princesses be passive and salvaged simply with a kiss," they note the reflection of the character in later Disney Princess animated films and in woman warriors such as Ellen Ripley from the Alien franchise and Xena of the adventure TV series Xena: Warrior Princess. A. O. Scott of The New York Times described Leia as "a foremother of Hermione Granger and Katniss Everdeen and of countless latter-day Disney princesses. She also foretold the recent, somewhat belated feminist turn in the Star Wars cycle itself".
Mark Hamill described Fisher's performance as:
Fisher herself described Leia as a "huge" feminist icon, dismissing the suggestion that the character was ever a "damsel in distress". Fisher said of Leia, "She bossed them around. I don't know what your idea of distress is, but that wasn't it! And I wasn't some babe running through the galaxy with my tits bouncing around. So I wasn't threatening to women". She added, "I like Princess Leia. I like how she was feisty. I like how she killed Jabba the Hutt". "I think I am Princess Leia, and Princess Leia is me. It's like a Möbius striptease."
Metal bikini
Leia's slave costume when she is held captive by Jabba the Hutt at the beginning of Return of the Jedi—made of brass and dubbed Leia's "Metal Bikini" or "Gold Bikini"—immediately made the character (and Fisher) a "generational sex symbol" celebrated by pin-up posters, and later merchandising and cosplay. The outfit has gained a cult following of its own.
Rosenberg noted that "the costume has become culturally iconic in a way that has slipped loose from the context of the scenes in which Leia wore it and the things she does after she is forced into the outfit." Wired wrote in 2006, "There's no doubt that the sight of Carrie Fisher in the gold sci-fi swimsuit was burned into the sweaty subconscious of a generation of fanboys hitting puberty in the spring of 1983." Acknowledging the opinion of some that the "Slave Leia" iconography tarnishes the character's position as "feminist hero", Rosenberg argues:
Science fiction filmmaker Letia Clouston concurs, saying "Sci-fi has had a long history of strong female characters. Yes, Princess Leia was in a gold bikini, but she was also the one who single-handedly killed Jabba. When you take into account movies and TV shows like Terminator, Aliens, Battlestar Galactica, and even video games like Metroid, you can see sci-fi has consistently promoted the strength of women more than any other genre."
Relationships
Family tree
Mentorship tree
See also
List of Kenner Star Wars action figures
Solo family
References
Footnotes
Citations
External links
Leia Organa on IMDb
Characters created by George Lucas
Female characters in film
Adoptee characters in films
Fictional ambassadors
Film characters introduced in 1977
Fictional commanders
Fictional diplomats
Fictional feminists and women's rights activists
Fictional female generals
Fictional women soldiers and warriors
Fictional ghosts
Fictional heads of state
Fictional military strategists
Fictional extraterrestrial princesses
Fictional revolutionaries
Fictional senators
Fictional twins
Fictional war veterans
Star Wars Anthology characters
Star Wars comics characters
Star Wars Jedi characters
Star Wars literary characters
Star Wars Skywalker Saga characters
Star Wars television characters
Star Wars video game characters
Fictional princesses |
null | null | Star Wars: Episode I – The Phantom Menace | eng_Latn | Star Wars: Episode I – The Phantom Menace is a 1999 American epic space opera film written and directed by George Lucas, produced by Lucasfilm, distributed by 20th Century Fox and starring Liam Neeson, Ewan McGregor, Natalie Portman, Jake Lloyd, Ian McDiarmid, Anthony Daniels, Kenny Baker, Pernilla August, and Frank Oz. It is the fourth film in the Star Wars film series and first chronological chapter of the "Skywalker Saga". Set 32 years before the original trilogy, during the era of the Galactic Republic, the plot follows Jedi Master Qui-Gon Jinn and his apprentice Obi-Wan Kenobi as they try to protect Queen Padmé Amidala of Naboo in hopes of securing a peaceful end to an interplanetary trade dispute. Joined by Anakin Skywalker—a young slave with unusually strong natural powers of the Force—they simultaneously contend with the mysterious return of the Sith.
Following the release of Return of the Jedi, Lucas was unmotivated to return to the franchise and continue the story beyond Return of the Jedi, though the backstory he created on Anakin sparked interest in him to develop a prequel trilogy. After he determined that computer-generated imagery (CGI) had advanced to the level he wanted for the prequel trilogy's visual effects, Lucas began writing The Phantom Menace in 1993 and production began in 1994. Filming started on June 26, 1997, at locations including Leavesden Film Studios and the Tunisian desert and ended on September 30. The film was Lucas's first directorial effort after a 22-year hiatus following the original Star Wars in 1977.
The Phantom Menace was released in theaters on May 19, 1999, almost 16 years after the premiere of Return of the Jedi. The film's premiere was extensively covered by media and was greatly anticipated because of the large cultural following the Star Wars saga had cultivated. Upon its release, The Phantom Menace received mixed reviews. While the visual effects, action sequences, musical score, and some performances (particularly Neeson and McGregor) were praised, criticism was largely focused on the screenplay, pacing, and characters (particularly Jar Jar Binks). Despite the reception, The Phantom Menace was a box office success and broke numerous box office records during its debut. It grossed more than $924.3 million worldwide during its initial theatrical run, becoming the highest-grossing film of 1999, the second-highest-grossing film worldwide and in North America (behind Titanic), and the highest-grossing Star Wars film at the time. A 3D reissue, which earned an additional $102.7 million and brought the film's overall worldwide takings to over $1 billion, was released in February 2012. Attack of the Clones (2002) and Revenge of the Sith (2005) followed The Phantom Menace, rounding out the Star Wars prequel trilogy.
Plot
The Trade Federation creates turmoil in the Galactic Republic by blockading the planet Naboo in protest of recent legislation taxing major galactic trade routes. The Republic's leader, Supreme Chancellor Finis Valorum, dispatches Jedi Master Qui-Gon Jinn and his apprentice, Obi-Wan Kenobi, to negotiate with Trade Federation Viceroy Nute Gunray. Darth Sidious, a Sith Lord and the Trade Federation's secret benefactor, orders the Viceroy to kill the Jedi and begin an invasion with an army of battle droids, but the Jedi escape and flee to Naboo. During the invasion, rescues a Gungan outcast, Jar Jar Binks. Indebted to , leads the Jedi to Otoh Gunga, the Gungans' underwater city. The Jedi fail to persuade the Gungan leader, Boss Nass, to help the planet's surface dwellers, but manage to obtain guidance and underwater transport to Theed, the capital city of Naboo. After rescuing Queen Amidala, the group make their escape from Naboo aboard her Royal Starship, intending to reach the Republic capital planet of Coruscant.
Passing through the Federation blockade, the ship is damaged in the crossfire, and its hyperdrive malfunctions. The ship lands for repairs on the outlying desert planet of Tatooine, situated beyond the Republic's jurisdiction. Qui-Gon, Jar Jar, astromech droid R2-D2, and Queen Padmé Amidala (in disguise as one of her own handmaidens) visit the settlement of Mos Espa to purchase a new part for their hyperdrive. They encounter a junk dealer, Watto, and his nine-year-old slave, Anakin Skywalker, a gifted pilot and engineer who has built a protocol droid, C-3PO.
Qui-Gon senses a strong presence of the Force within Anakin, and is convinced that he is the prophesied "Chosen One," destined to restore balance to the Force. With Watto refusing to accept payment in Republic currency, wagers both the required hyperdrive part and Anakin's freedom in a podrace. Anakin wins the race and joins the group to be trained as a Jedi, leaving behind his mother, Shmi. En route to their starship, encounters Darth Maul, Sidious' apprentice, who was sent to capture Amidala. After a brief lightsaber duel, escapes aboard the starship with the others.
Qui-Gon and Obi-Wan escort Amidala to Coruscant so that she can plead her people's case to Chancellor Valorum and the Galactic Senate. Qui-Gon informs the Jedi Council that his attacker was a Sith and subsequently asks for permission to train Anakin as a Jedi, but the Council refuses his request, concerned that Anakin is vulnerable to the dark side of the Force. Undaunted, vows to take up Anakin as his new apprentice. Meanwhile, Naboo's Senator Palpatine persuades Amidala to call for a vote of no confidence in Valorum to elect a more capable leader and to resolve the crisis. Though she is successful in pushing for the vote, Amidala grows frustrated with the corruption in the Senate, and decides to return to Naboo. and are ordered by the Jedi Council to accompany the queen and investigate the return of the Sith, whom they had thought to be extinct for over a millennium.
On Naboo, Padmé reveals herself as the actual queen before the Gungans to gain their trust, and persuades them to help against the Trade Federation. is promoted to general and joins his tribe in a battle against the droid army, while Padmé leads the search for Gunray in Theed. Qui-Gon tells Anakin to hide inside a starfighter in the palace hangar, but he accidentally triggers its autopilot, and flies into space, joining the Naboo pilots in their battle against the Federation droid control ship. With R2's help, Anakin pilots the fighter into the ship and causes its destruction from within, deactivating the droid army. Meanwhile, Maul, who has been dispatched by Sidious to assist Gunray, engages in a lightsaber duel with and . Maul mortally wounds Qui-Gon, but is then sliced in half by Obi-Wan and falls down a shaft. Qui-Gon asks Obi-Wan to train Anakin before dying in his arms.
Following the battle, Gunray is arrested by the Republic, and Palpatine is elected Chancellor. Master Yoda promotes to the rank of Jedi Knight, and reluctantly accepts Anakin as apprentice. A funeral is held for Qui-Gon, attended by the other Jedi, who contemplate that there is still one Sith remaining since there are always two of them. During a celebratory parade on Naboo, Padmé presents a gift of thanks to the Gungans to establish peace.
Cast
Liam Neeson as Qui-Gon Jinn, a Jedi Master who discovers Anakin and under the belief that he is destined to bring balance to the force as the prophesied "Chosen One", insists that the boy be trained as a Jedi, despite the Jedi Council's refusal to do so. Lucas originally wanted to cast an American actor in the role, but cast Neeson (who is Northern Irish) because he considered that Neeson had great skills and presence. Lucas said Neeson was a "master actor, who the other actors will look up to, who has got the qualities of strength that the character demands."
Ewan McGregor as Obi-Wan Kenobi, Qui-Gon's twenty-five-year-old Jedi Padawan, who holds his master in high regard but questions his motives at times, especially when it seems he is breaking the rules of the Jedi. McGregor was cast from a shortlist of fifty actors, all of whom had to be compared to pictures of young Alec Guinness, who portrayed the elderly Obi-Wan, to make a believable younger version. McGregor had a vocal coach to help his voice sound closer to Guinness's. He also studied several of Guinness's performances, from his early work and the original Star Wars films.
Natalie Portman as Queen Padmé Amidala, the fourteen-year-old Queen of Naboo, who hopes to protect her planet from the Trade Federation's blockade invasion. Throughout most of the film, she uses her birth name Padmé Naberrie and poses as one of the queen's handmaidens for protection. Over 200 actresses auditioned for the role. The production notes stated that "The role required a young woman who could be believable as the ruler of that planet, but at the same time be vulnerable and open". Portman was chosen especially for her performances in Léon: The Professional (1994) and Beautiful Girls (1996), which impressed Lucas.<ref name="ProductionNotes">{{cite web |url=https://www.starwars.com/episode-i/bts/production/f19990501/indexp4.html |archive-url=https://web.archive.org/web/20041023023458/http://www.starwars.com/episode-i/bts/production/f19990501/indexp4.html |archive-date=October 23, 2004 |date=May 1, 1999 |title=Star Wars Episode I Production Notes: The Actors and Characters – Part I |publisher=StarWars.com |access-date=April 5, 2009}}</ref> He stated, "I was looking for someone who was young, strong, along the lines of Leia [and] Natalie embodied all those traits and more". Portman was unfamiliar with Star Wars before being cast, but was enthusiastic about being cast as a character she expected to become a role model. Portman said, "It was wonderful playing a young queen with so much power. I think it will be good for young women to see a strong woman of action who is also smart and a leader."
Jake Lloyd as Anakin Skywalker, a nine-year-old slave boy and a skilled pilot who dreams of becoming a Jedi. Hundreds of actors were tested across the UK, Ireland, Canada, and the United States before the producers settled on Lloyd, who Lucas considered met his requirements of "a good actor, enthusiastic and very energetic". Producer Rick McCallum said that Lloyd was "smart, mischievous, and loves anything mechanicaljust like Anakin."
Ian McDiarmid as Senator Palpatine / Darth Sidious, a middle aged senator of Naboo and a secret Sith Lord who orchestrates the invasion of his home planet to get elected Supreme Chancellor. He is the Trade Federation's mysterious benefactor, and the titular "Phantom Menace". McDiarmid was surprised when Lucas approached him sixteen years after Return of the Jedi to reprise the role of Palpatine, having assumed that a younger actor would play the role in the prequel films.
Ahmed Best as Jar Jar Binks, a clumsy Gungan exiled from his home and taken in by Qui-Gon and Obi-Wan. Best was hired after casting director Robin Gurland saw him on a Stomp performance in San Francisco. Best was originally intended to provide motion capture data, but his offer to voice the character was accepted. On the set, to provide references for the actors, Best was clothed in a suit made of foam and latex and a headpiece. Best's filmed performance was later replaced with the computer-generated character. Best frequently improvised movements to make Jar Jar look as clumsy and comedic as possible.
Anthony Daniels as the voice of C-3PO, a protocol droid built by Anakin. He lacks a metal covering in this film; R2-D2 humorously refers to it as being "naked". A puppeteer dressed in a color closely matching the backgroundin a manner similar to the Japanese puppet theater Bunrakumanipulated a skeletal C-3PO figure attached to his front while Daniels read his lines off-camera. The puppeteer was erased from the film during post-production.
Kenny Baker as R2-D2, an astromech droid from Naboo that saves Queen Amidala's ship when other astro droids fail. Before the film's production started, fans campaigned on the Internet to retain Baker as R2-D2; Lucas replied that the actor would reprise the role. Baker is used for scenes where R2-D2 bends forwards and backwards and wobbles from side-to-side. Robots and a digital model were used in other shots.
Pernilla August as Shmi Skywalker, Anakin's mother, who is concerned for her son's future and allows him to leave with the Jedi. August, a veteran of Swedish cinema, was chosen after auditioning with Liam Neeson. She was afraid of being rejected because of her accent.
Frank Oz as the voice of Yoda, the centuries-old Jedi Grandmaster and head of the Jedi Council who is apprehensive about allowing Anakin to be trained. Yoda was mostly portrayed as a puppet designed by Nick Dudman based on Stuart Freeborn's original design. Oz controlled the puppet's mouth, and other parts were controlled by puppeteers using remote controls. Lucas fitted Yoda's filming around Oz's schedule as he finished and promoted In & Out. A computer-generated Yoda is featured in two distant shots. Warwick Davis (who played the part of the Ewok Wicket W. Warrick in Return of the Jedi in 1983) portrays him in the scene where Obi-Wan becomes a Jedi Knight. Lucas said he originally wanted to use a full-time digital Yoda, but the attempts did not work well enough at the time. Beginning with the 2011 Blu-ray release of The Phantom Menace, which was also used for the 3D reissue, a CG Yoda replaced the puppet entirely.
Oliver Ford Davies as Sio Bibble, the governor of Naboo.
Hugh Quarshie as Captain Quarsh Panaka, Queen Amidala's chief of security at the Theed Palace.
Samuel L. Jackson as Mace Windu, a Jedi Master and high-ranking member of the Jedi Council who opposes training Anakin.
Ray Park as Darth Maul, Darth Sidious' Zabrak Sith apprentice, who wields a double-bladed lightsaber.
Peter Serafinowicz as the voice of Darth Maul
Terence Stamp as Chancellor Finis Valorum, the Supreme Chancellor of the Galactic Republic who commissions Obi-Wan and Qui-Gon to negotiate with the Trade Federation viceroy. Lucas described the character as a "good man but he's beleaguereda bit like [Bill] Clinton".
Keira Knightley as Sabé, one of Queen Amidala's handmaidens who serves as her decoy throughout the majority of the film.
Silas Carson as:
Nute Gunray, the viceroy of the Trade Federation who leads Naboo's invasion and tries to force Queen Amidala to sign a treaty to legitimize their occupation of the planet.
Ki-Adi-Mundi, a wise and powerful Jedi Master who sits on the Council.
Lott Dod, a Trade Federation Senator.
An ill-fated pilot. This was the role for which Carson originally auditioned.
Jerome St. John Blake as:
Rune Haako, Gunray's chief lieutenant and Settlement Officer in the Trade Federation.
James Taylor as the voice of Rune Haako.
Oppo Rancisis, a Jedi Master and member of the Council.
Orn Free Ta, a Twi-lek senator.
Mas Amedda, a Chagrian politician and Vice Chair of the galactic senate.
Additionally, Brian Blessed, Andy Secombe, and Lewis MacLeod voiced Boss Nass, the leader of the Gungan tribe who allies with the Naboo surface dwellers, Watto, a junk dealer on Tatooine who owns Anakin and his mother as slaves, and Sebulba, an aggressive, scheming podracer who is Anakin's main rival at the Boonta Eve podrace, respectively. Greg Proops and Scott Capurro voiced Fode and Beed, respectively, the two-headed announcer of the Boonta Eve Race. Alan Ruscoe appears as Jedi Master Plo Koon and Neimoidian Daultay Dofine, commander of the Trade Federation's droid control ships. Ralph Brown plays Ric Olie, commander of the Naboo Royal Space Fighter Corps and chief pilot aboard Queen Amidala's starship, Matthew Wood appears as the Twi'lek Bib Fortuna, alongside a CGI Jabba the Hutt who is voiced by an uncredited actor. Dominic West plays the role of an unnamed Naboo guard, and Sofia Coppola appears as Saché, one of Amidala's handmaidens. Christian Simpson appears as Lieutenant Gavyn Sykes. Lindsay Duncan voices TC-14, a protocol droid on the Federation ship. Sally Hawkins made her screen debut as an uncredited villager.
Production
Development
While writing the original Star Wars film, Lucas decided the story was too vast to be covered in one film. He introduced a wider story arc that could be told in sequels if it became successful. He negotiated a contract that allowed him to make two sequels, and over time created an elaborate backstory to aid his writing process. While writing the second film, The Empire Strikes Back, Lucas considered directions in which to take the story. In the original trilogy, Darth Vader was revealed to have been Anakin Skywalker, a once-powerful Jedi Knight, and a traitor to the Jedi Order. With this backstory in place, Lucas decided that the movies would work best as a trilogy. In the trilogy's final episode, Return of the Jedi, Vader is redeemed through an act of sacrifice for Luke.
Throughout the 1980s, Lucas said he had no desire to return to Star Wars and had canceled his sequel trilogy by the time of Return of the Jedi. However, because Lucas had developed most of the backstory, the idea of prequels continued to fascinate him. In the early 1990s, Star Wars saw a resurgence in popularity in the wake of Dark Horse's comic line and Timothy Zahn's trilogy of novels. Lucas saw that there was still a large audience for his idea of a prequel trilogy, and with the development of special effects generated with computer-generated imagery (CGI), Lucas considered returning to his saga and directing the film. In 1993, it was announced in Variety and other sources that he would be making the prequels. Lucas began outlining the story; Anakin Skywalker rather than Obi-Wan Kenobi would be the protagonist, and the series would be a tragedy examining Darth Vader's origins. A relic of the original outline was that Anakin would, like his son, grow up on Tatooine. Lucas also began to change the prequels' timeline relative to the original series; instead of filling in the tangential history, they would form the beginning of a long story that started with Anakin's childhood and ended with his death. This was the final step toward turning the franchise into a saga.
Lucas began writing the Star Wars prequel trilogy on November 1, 1994. The screenplay of Star Wars was adapted from Lucas' 15-page outline that was written in 1976, which he designed to help him keep track of the characters' backstories and events that occurred before the original trilogy. Anakin was first written as a twelve-year-old, but Lucas reduced his age to nine because he felt that the lower age would better fit the plot point of Anakin being affected by his mother's separation from him. Eventually, Anakin's younger age led Lucas to rewrite his participation in the movie's major scenes. The film's working title was The Beginning, with the title not being changed to The Phantom Menace until shortly before the film's completion. Lucas later revealed that the Phantom Menace title was a reference to Palpatine hiding his true identity as an evil Sith Lord behind the facade of a well-intentioned public servant.
The larger budget and possibilities opened up by the use of digital effects made Lucas "think about a much grander, more epic scale"which is what I wanted Star Wars to be". The story ended with five simultaneous, ongoing plots, one leading to another. The central plot is Palpatine's intent to become Chancellor, which leads to the Trade Federation's attack on Naboo, the Jedi being sent there, Anakin being met along the way, and the rise of the Sith Lords. As with the original trilogy, Lucas intended The Phantom Menace to illustrate several themes throughout the narrative. Duality is a frequent theme; Amidala is a queen who passes as a handmaiden, Palpatine plays on both sides of the war, among others. "Balance" is frequently suggested; Anakin is supposedly "the one" chosen to bring balance to the ForceLucas said, "Anakin needed to have a mother, Obi-Wan needed a Master, Darth Sidious needed an apprentice" as without interaction and dialogue, "you wouldn't have drama".
In November 2015, Ron Howard confirmed that he, Robert Zemeckis, and Steven Spielberg were approached by Lucas to direct The Phantom Menace. All three approached directors told Lucas that he should direct the film, as they each found the project "too daunting."
Pre-production and design
Before Lucas had started writing, his producing partner Rick McCallum was preparing for the film. McCallum stated that his experience with The Young Indiana Jones Chronicles led to many of his decisions on The Phantom Menace, such as long-term deals with actors and soundstages, the employment of recent graduates with no film experience, and the creation of sets and landscapes with digital technology. In April 1994, McCallum started searching for artists in art, architecture and design schools, and in mid-year he began location scouting with production designer Gavin Bocquet. Industrial Light & Magic (ILM) art director Doug Chiang impressed McCallum the most and was hired as the design director. Art development on the film began in January 1995.
Within three to four months of Lucas beginning the writing process, Chiang and his design team started a two-year process of reviewing thousands of designs for the film. Chiang stated that Lucas intended Episode I to be stylistically different from the other Star Wars films; it would be "richer and more like a period piece, since it was the history leading up to A New Hope." The story takes place on three planets, some with varied environments such as the human and Gungan cities of Naboo and three buildings in Coruscant. With the exception of the Gungan city, which had an art nouveau-inspired visual, these locations would be given distinctive looks with some basis in the real world. The concept drawings of Ralph McQuarrie for the original trilogy served as the basis for Mos Espawhich was also inspired by old Tunisian hotels and buildings and had touches such as a market place to differentiate it from A New Hopes Mos Eisleyand Coruscant, in particular a metropolis design that became the basis for the Senate. Bocquet would later develop the work of Chiang's team and design the interiors, translating the concepts into construction blueprints with environments and architectural styles that had some basis in reality "to give the audience something to key into." Some elements were directly inspired by the original trilogy; Lucas described the battle droids as predecessors to the stormtroopers. Chiang uses that orientation to base the droids on the Imperial soldiers, only in the same style of stylized and elongated features seen in tribal African art.
Terryl Whitlatch, who had a background on zoology and anatomy, was in charge of creature design. Many of the aliens are hybrids, combining features of real animals. At times, entire food chains were developed, even though only a small percentage of them would appear in the film. Whitlatch also designed detailed skeletons for the major characters and facial muscles on Jar Jar Binks as a reference for ILM's animators. Each creature would reflect its environment; those on Naboo were more beautiful because the planet is "lush and more animal-friendly", Tatooine has rough-looking creatures "with weather-beaten leathery skin to protect them from the harsh desert elements", and Coruscant has bipedal, human-looking aliens.
The film made extensive use of the new technique of digital pre-visualization, using computers to essentially create 3-D animated storyboards. This was done for dozens of scenes in the film but was first and primarily used in the pod race sequence. Animatic supervisor David Dozoretz, also an ILM alum, worked on this sequence for nearly three years, and at one point had a 25 minute version of the race, although the film only included a 9-minute version.
Stunt coordinator Nick Gillard was recruited to create a new Jedi fighting style for the prequel trilogy. Gillard likened the lightsaber battles to a chess game "with every move being a check". Because of their short-range weapons, Gillard thought that the Jedi would have had to develop a fighting style that merged every swordfighting style, such as kendo and other kenjutsu styles, with other swinging techniques, such as tennis swings and tree-chopping. While training Liam Neeson and Ewan McGregor, Gillard wrote a sequence that lasted around 60 seconds and intended to be around five or six sequences per fight. Lucas later referred to the Jedi as "negotiators" rather than high-casualty soldiers. The preference of hand-to-hand combat was intended to give a spiritual and intellectual role to the Jedi. Because Gillard thought that the stunt jumps with the actors and stuntmen dangling from wires did not look realistic, air rams were used to propel them into the air instead.
Lucas decided to make elaborate costumes, because the film's society was more sophisticated than the one depicted in the original trilogy. Designer Trisha Biggar and her team created over 1,000 costumes that were inspired by various cultures. Biggar worked closely with concept designer Iain McCaig to create a color palette for the inhabitants of each world: Tatooine followed A New Hope with sun-bleached sand colors, Coruscant had grays, browns, and blacks, and Naboo had green and gold for humans while Gungans wore "a leathery look, like their skin". The Jedi costumes followed the tradition from the original film; Obi-Wan's costume was inspired by the costume that was worn by Guinness. Lucas said he and Biggar would look at the conceptual art to "translat[e] all of these designs into cloth and fabric and materials that would actually work and not look silly". Biggar also consulted Gillard to ensure that the costumes would accommodate action scenes, and consulted the creature department to find which fabrics "wouldn't wear too heavily" on the alien skins. A huge wardrobe department was set up at Leavesden Film Studios to create over 250 costumes for the main actors and 5,000 for the background ones.
Nute Gunray's Thai accent was chosen after Lucas and McCallum listened to various languages to decide how the Neimoidians would speak. The character design of Watto was an amalgam of rejected ideas; his expressions were based on video footage of Secombe's voice acting, photographs of animation supervisor Rob Coleman imitating the character, and modeler Steve Alpin saying Watto's lines to a mirror. Lucas described Sebulba's design as "a spider crossed with an orangutan crossed with a sloth", with a camel-like face, and clothing inspired by medieval armor.
Casting
After Samuel L. Jackson expressed interest in appearing in a Star Wars film, he was approached by casting director Robin Gurland to play Windu. Tupac Shakur was also considered for the role of Mace Windu. Ray Park, a martial arts champion with experience in gymnastics and sword fighting, was originally a member of the stunt crew. Stunt coordinator Nick Gillard filmed Park to demonstrate his conception of the lightsaber battles. Lucas and McCallum were so impressed with the test tape that they gave Park the role of Maul. His voice was considered "too squeaky" and was dubbed over in post-production by Peter Serafinowicz. Keira Knightley's parents tried to convince her not to audition, but the teenage actress still sought a role since she was a Star Wars fan. The casting was influenced by Knightley's remarkable similarity to Natalie Portman, with the actress admitting their mothers could not tell each other apart. Knightley was reported to have "cried every single day" due to finding the wardrobe uncomfortable.
Vinette Robinson auditioned for the role of Padmé Amidala.
Over 3,000 actors auditioned for the role of Anakin Skywalker including Cameron Finley, Justin Berfield and Michael Angarano.
Benicio del Toro was originally cast as Darth Maul but later left the project when the character’s lines were cut.
Michael Jackson expressed interest in playing Jar Jar Binks but he wanted to do it in prosthetic make ups while George Lucas wanted to do it in CGI.
Joseph Fiennes auditioned for the role of Obi-Wan Kenobi and nearly landed the part until George Lucas's young daughter rejected him upon meeting him during the second level of auditioning.
Silas Carson was cast as Nute Gunray because another actor was uncomfortable with the costumes used by the Trade Federation characters, which were hot, exerted a lot of pressure on the bearer, and took about 15 minutes to apply. Hugh Quarshie considered the part of Panaka as "a good career move" and a production that would be fun to make. Brian Blessed originally auditioned for the role of Sio Bibble, the Governor of Naboo, for which he was considered "too loud". Casting director Robin Gurland approached him to play Nass because it was a "bigger than life" character with "a kind of bravado". Blessed described Nass as a "reluctant hero". Sofia Coppola, daughter of Lucas' long-time friend and creative partner Francis Ford Coppola, considers Lucas as "like an uncle to me". As she prepared the script for her directorial debut The Virgin Suicides, Coppola heard Lucas would make a new Star Wars film and asked him if she could accompany him during filming. Lucas offered Coppola a role in the royal entourage, which she accepted because it "seemed like a good vantage point to watch without getting in the way".
Filming
Filming began on June 26, 1997, and ended on September 30 of that year, primarily taking place at Leavesden Film Studios in England. Leavesden was leased for a two and a half year period so the production company could leave the sets intact and return after principal photography had been completed. The forest scenes on Naboo were filmed at Cassiobury Park in Watford, Hertfordshire. Pick-ups were shot between August 1998 and February 1999 after Lucas screened a rough cut of the film for friends and colleagues in May 1998. Most of the action and stunts were filmed by Roger Christian's second unit, which worked alongside the main unit instead of afterwards because of the high number of shots to be completed daily.
The Tunisian desert was again used for the Tatooine scenes; Mos Espa was built outside the city of Tozeur. On the night following the third day of shooting in Tozeur, an unexpected sandstorm destroyed many of the sets and props. The production was quickly rescheduled to allow for repairs and was able to leave Tunisia on the date originally planned. The Italian Caserta Palace was used as the interior of the Theed City Naboo Palace; it was used as a location for four days after it had been closed to visitors. Scenes with explosions were filmed on replica sets in Leavesden.
A binder with the film's storyboards served as a reference for live-action filming, shots that would be filmed in front of a chroma key blue screen, and shots that would be composed using CGI. The sets were often built with the parts that would be required on screen; often they were built only up to the heights of the actors. Chroma key was extensively used for digital set extensions, backgrounds, or scenes that required cinematographer David Tattersall to seek powerful lamps to light the sets and visual effects supervisor John Knoll to develop software that would remove the blue reflection from shiny floors. Knoll, who remained on set through most of the production, worked closely with Tattersall to ensure that the shots were suitable to add effects later. The cameras were fitted with data capture models to provide technical data for the CGI artists.The Phantom Menace was the final Star Wars film to be shot on 35mm film until Episode VII (Star Wars: The Force Awakens). Some scenes, mostly of elements filmed by the special effects team, were shot on high definition, digital video tapes to test the performance of digital recordings, which Lucas and McCallum considered the next logical step because of the amount of digitizingan expensive process compared to recording directly on digital mediafor the compositing of computer-generated effects. All future films would be shot using Sony CineAlta high-definition video cameras. Greg Proops and Scott Capurro were filmed wearing makeup and blue bodysuits so their heads could be joined in a computer-generated body. The visual effects crew did not like the original results and crafted Fode and Beed as an entirely computer generated alien.
Editing took two years; Paul Martin Smith started the process in England and focused on dialogue-heavy scenes. Ben Burttwho was also the film's sound editorwas responsible for action sequences under Lucas' supervision. Non-linear editing systems played a large part in translating Lucas' vision; he constantly tweaked, revised, and reworked shots and scenes. The final sound mix was added in March 1999, and the following month, the film was completed after the delivery of the remaining visual effects shots.
Effects
The film saw breakthrough in computer generated effects. About 1,950 of the shots in The Phantom Menace have visual effects. The scene in which toxic gas is released on the Jedi is the only sequence with no digital alteration. The work was so extensive that three visual effects supervisors divided the workload among themselvesJohn Knoll supervised the on-set production and the podrace and space battle sequences, Dennis Muren supervised the underwater sequence and the ground battle, and Scott Squires, alongside teams assigned for miniature effects and character animation, worked on the lightsaber effects.
Until the film's production, many special effects in the film industry were achieved using miniature models, matte paintings, and on-set visual effectsalthough other films had made extensive use of CGI. Knoll previewed 3,500 storyboards for the film; Lucas accompanied him to explain factors of the shots that would be practical and those that would be created through visual effects. Knoll later said that on hearing the explanations of the storyboards, he did not know how to accomplish what he had seen. The result was a mixture of original techniques and the newest digital techniques to make it difficult for the viewer to guess which technique was being used. Knoll and his visual effects team wrote new computer software, including cloth simulators to allow a realistic depiction of the digital characters' clothing, to create certain shots. Another goal was to create computer-generated characters that could act seamlessly with live-action actors. While filming scenes with CGI characters, Lucas would block the characters using their corresponding voice actors on-set. The voice actors were then removed and the live-action actors would perform the same scene alone. A CGI character would later be added into the shot to complete the conversation. Lucas also used CGI to correct the physical presence of actors in certain scenes. Practical models were used when their visuals helped with miniature sceneries for backgrounds, set extensions, and model vehicles that would be scanned to create the digital models or filmed to represent spaceships and podraces.
Lucas, who had previously confronted problems with the props used to depict R2-D2, allowed ILM and the production's British special effects department to create their own versions of the robot. Nine R2-D2 models were created; one was for actor Kenny Baker to be dropped into, seven were built by ILM and featured two wheelchair motors capable of moving , enabling it to run and be mostly used in stage sets, and the British studio produced a pneumatic R2-D2 that could shift from two to three legs and was mostly used in Tunisia because its motor drive system allowed it to drive over sand.
Lucas originally planned to create many of the aliens with computer graphics, but those that would be more cost-effectively realized with masks and animatronics were created by Nick Dudman's creature effects team. These included the Neimoidians, background characters in Mos Espa, the Jedi Council, and the Galactic Senate. Dudman's team was told where the creatures would be required six months before principal photography begun, and they rushed the production. The Neimodian suits, which were originally intended as digital characters, were delivered one day before they would be required on set. Dudman traveled to Skywalker Ranch to see the original creatures that could be reused, and read the script for a breakdown of scenes with practical creatures, leaving only the more outlandish designs to be created using CGI.
To research for the podrace vehicles, the visual effects crew visited a jet aircraft junkyard outside Phoenix, Arizona and scavenged four Boeing 747 engines. Life-sized replicas of the engines were built and sent to Tunisia to provide reference in the film. Except for Jake Lloyd inside a hydraulically controlled cockpit and a few practical podracer models, the entire podracing scenewhich the effects crew designed to be as "out of this world" as possibleis computer-generated.
Themes
Like previous Star Wars films, The Phantom Menace makes several references to historical events and films that George Lucas watched in his youth. The Star Wars films typically mix several concepts from different mythologies together, drawing heavily from the hero's journey, an archetypical template developed by comparative mythologist Joseph Campbell. This film also notably borrows from Ben Hur, including the podrace sequence, which mirrors the chariot race scene; additionally, the end celebration scene closely resembles a Roman parade in Ben Hur.
There are many references to Christianity in the film, such as the appearance of Darth Maul, whose design draws heavily from traditional depictions of the devil, complete with red skin and horns. The Star Wars film cycle features a similar Christian narrative involving Anakin Skywalker; he is the "Chosen One"—the individual prophesied to bring balance to the Force—who was conceived of a virgin birth. However, unlike Jesus, Anakin will eventually fall from grace and seemingly fail to fulfill his destiny (until the prophecy comes true in Return of the Jedi).
Japanese films such as Akira Kurosawa's The Hidden Fortress influenced the original Star Wars film; scholars say that The Phantom Menace was likewise influenced by Korean and Japanese culture. Film historians Geoff King and Tanya Krzywinska write, "The costume and make-up designs ... favour a mixture of the gothic and the oriental over anything very futuristic. The gothic is most strongly apparent in Darth Maul's demonic horns and the red and black make-up mask that borrows from the facial designs found in depictions of Japanese demons". King and Krzywinska say that "Qui-Gon's pony tail and Obi-Wan's position of apprentice further encourage a reading in terms of the Samurai tradition". They also say "Amidala, in keeping with her status and character, has a number of highly formal outfits ... to go with hair sculpted into a curve that frames make-up of a Japanese cast".
The Jedi practice Zen-like meditation and martial arts, as did the ancient Japanese Samurai warriors. The name "Qui-Gon" adapts the term qigong, which refers to a Chinese discipline involving meditation and cultivation of the flow of the vital energy called "Chi" or "Qi" for healing, health, and combat. The words Ch'i (Chinese), gi (Korean), ki (Japanese), and the Indian term "Prana" all refer to the energy that is thought to flow through all living things, from the source of all chi (or power), which is "The Way" or "The Tao" in Chinese philosophy. In Taoist philosophy, from The Way, yin and yang—the opposing but complementary aspects of reality or nature—are born. Unlike Chinese philosophy, in which yin and yang are not moral qualities, the ancient Persian philosophy of Zurvanism taught that the dualism of dark and light forces are locked in an eternal battle while being two sides (or evolutes) of the same "Force", the force of time itself (Zurvan)—the prime mover. These elements derive primarily from Eastern and Iranian religions and myths.
As with other Star Wars films, the themes about family and hope are featured prominently. In the episode "Legacy" of Disney Gallery: The Mandalorian, Dave Filoni explains that the use of "Duel of the Fates" during the lightsaber duel between Darth Maul against Qui-Gon Jinn and Obi-Wan Kenobi represents the fight for the fate of Anakin Skywalker. According to Filoni, Qui-Gon Jinn acts like a father figure towards Anakin because he feels he needs one after taking him from his mother, having realized that the Jedi shouldn't be opposed to love and care. In the end, Qui-Gon dies, and thus Anakin loses his father figure; Obi-Wan Kenobi ultimately becomes his master to honor Jinn's dying wish despite his initial contempt for him, and while he comes to eventually see him like a brother as shown in Attack of the Clones and Revenge of the Sith, he doesn't act like a father, which coupled with the Jedi's indifference, seals Anakin's fate.
Music
As with previous Star Wars films, Star Wars Episode I: The Phantom Menaces score was composed and conducted by John Williams. He started composing the score in October 1998 and began recording the music with the London Voices and London Symphony Orchestra at Abbey Road Studios on February 10, 1999. Williams decided to use electronic instruments such as synthesizers to enhance the sound and choral pieces to "capture the magical, mystical force that a regular orchestra might not have been able to provide", and create an atmosphere that was "more mysterious and mystical and less military" than those of the original trilogy. One of the most notable tracks is "Duel of the Fates", which uses the chorus to give a religious, temple-like feel to the epic lightsaber duel. The track was made into a music video. While composing Anakin's theme, Williams tried to reflect the innocence of his childhood and to foreshadow his transformation into Darth Vader by using slight suggestions of "The Imperial March" in the melody.
The film's soundtrack was released by Sony Classical Records on May 4, 1999. This album featured the score, which Williams restructured as a listening experience; it is not presented in film order and omits many notable cues from the film because of the space restriction of the compact disc. A two-disc "Ultimate Edition" was released on November 14, 2000. The set features the entire score as it is heard in the film, including all of the edits and loops that were made for the sound mix.
Marketing
Promotion
Lucasfilm spent on the film's advertising campaign and made promotional licensing deals with Hasbro, Lego, Tricon Global Restaurants, and PepsiCo. Lucasfilm also helped the Star Wars fan club to organize an event called Star Wars Celebration, which was held in Denver, Colorado between April 30 and May 2, 1999.
The teaser trailer was released on selected screens accompanying Meet Joe Black on November 13, 1998, and media reported that people were paying full admission at theaters to see the trailer. To keep fans from leaving before the movie was over, some theaters played the teaser an additional time after the film finished. A second trailer was released on March 12, 1999, with the film Wing Commander. Again, many fans paid full theater admission to watch the new trailer. A bootlegged version of the preview was leaked to the Internet the same day. The next morning, the trailer was released on the film's official website and shortly afterwards the servers became overloaded. The theatrical trailer caused even more media attention because it was premiered in theaters and screened at the ShoWest Convention in Las Vegas, and was aired on Entertainment Tonight and Access Hollywood.
The teaser poster, featuring Anakin with his shadow forming Darth Vader's silhouette, was released on November 10, 1998. After Lucas opted for a drawn theatrical poster, Drew Struzan, the artist responsible for the Special Edition posters, was commissioned to illustrate, and the poster was unveiled on March 11, 1999. Lucasfilm dictated that, contractually, Struzan's illustration was the only art the foreign distributors could use, and other than the text, it could not be modified in any way.
Adaptations
A novelization was written by Terry Brooks, who met with Lucas before writing the book and received his approval and guidance. It includes some scenes that are not in the film and foreshadows pending developments in the following two installments of the series. Additionally, a Scholastic junior novelization was written by Patricia C. Wrede. A four-issue comic book adaptation was written by Henry Gilroy and published by Dark Horse Comics. The film was again adapted as part of the all-ages Star Wars: The Prequel Trilogy - A Graphic Novel, released by Disney-Lucasfilm Press in 2017.
A LucasArts video game adaptation was released for the PlayStation and PC. The podracing tie-in Star Wars Episode I: Racer was released for Nintendo 64, PC, and other platforms. A pinball machine was created by WMS Industries.
Release
Theatrical
The release of the first new Star Wars film in 16 years was accompanied by a considerable amount of attention. According to The Wall Street Journal, so many workers announced plans to view the premiere that many companies closed on the opening day. Queue areas formed outside cinema theaters over a month before ticket sales began. More theater lines appeared when it was announced that cinemas were not allowed to sell tickets in advance until two weeks into the release. This was because of a fear that family theater-goers would be either unable to receive tickets or would be forced to pay higher prices for them. Instead, tickets were to be sold on a first-come-first-served basis. However, after meetings with the National Association of Theatre Owners, Lucasfilm agreed to allow advance ticket sales on May 12, 1999, provided there was a limit of 12 tickets per customer. As a result, some advance tickets were sold by scalpers at prices as high as apiece, which a distribution chief called "horrible" and said was exactly what they wanted to avoid. Daily Variety reported that theater owners received strict instructions from Lucasfilm that the film could only play in the cinema's largest auditorium for the first 8–12 weeks, no honor passes were allowed for the first eight weeks, and they were obliged to send their payments to distributor 20th Century Fox within seven days.
Despite worries about the film being finished on time, two weeks before its theatrical release Lucasfilm moved the release date from May 21 to 19, 1999. At the ShoWest Convention, Lucas said the change was intended to give the fans a "head start" by allowing them to view it during the week and allowing families to view it during weekends. Eleven charity premieres were staged across the United States on May 16, 1999; receipts from the Los Angeles event, where corporate packages were available for between and ; proceeds were donated to the Elizabeth Glaser Pediatric AIDS Foundation. Other charity premieres included the Dallas premiere for the Children's Medical Center, the Aubrey Fund for Pediatric Cancer Research at the Sloan-Kettering Hospital in New York, the Big Brother/Sister Association of the Philadelphia premiere, and the Children's National Medical Center in Washington, D.C. A statement said that tickets were sold at apiece and that certain sections of the theaters were set aside for disadvantaged children.
The film opened at 12:01 am on Wednesday, May 19, 1999 in 2,010 theaters in the United States and Canada. An additional 960 theaters screened the film later in the day. About 120 theaters showed the film continuously on opening day, including the Ziegfeld Theatre in New York City and the Chinese Theatre in Los Angeles. Foreshadowing his future conversion to digital cinematography, Lucas said the film would be released on four digital projectors (two in New York and two in Los Angeles) on June 18, 1999. Few film studios released films during the same week: DreamWorks and Universal Studios released The Love Letter on May 21 and Notting Hill on May 28, respectively. The Love Letter was a commercial failure but Notting Hill fared better and followed The Phantom Menace closely in second place. Employment consultant firm Challenger, Gray & Christmas estimated that 2.2 million full-time employees missed work to attend the film, resulting in a loss of productivity.
Home media
The film was released worldwide on VHS between April 3 and 8, 2000. Two versions were released in North America on April 4a standard pan and scan version and a widescreen Collector's Edition version. In its first two days of availability, the regular version sold 4.5 million copies and the limited edition sold 500,000. It was the first Star Wars film to be officially released on DVD, on October 16, 2001, in a slightly extended cut from the theatrical release. The special features included seven deleted scenes completed specifically for the DVD, a commentary track featuring Lucas and producer Rick McCallum, and several documentariesincluding a full-length documentary entitled The Beginning: Making Episode I. The Phantom Menace became the fastest selling DVD ever in the U.S.; 2.2 million copies were sold in its first week after release.
The DVD version was re-released in a prequel trilogy box set on November 4, 2008. A LaserDisc version of The Phantom Menace was released in Japan on April 7, 2000, one year and six months before it was available on DVD in the U.S. The Star Wars films were released by 20th Century Fox Home Entertainment on Blu-ray Disc on September 16, 2011; The Phantom Menace was restored to improve the picture quality and remove the magnification present on the previous DVD release, restoring approximately 8 percent of the picture to the frame. In the Blu-ray release of The Phantom Menace, the Yoda puppet was replaced with a CGI model, making it consistent with the other films of the prequel trilogy.
On April 7, 2015, Walt Disney Studios, 20th Century Fox, and Lucasfilm jointly announced the digital releases of the six released Star Wars films. The Phantom Menace was released through the iTunes Store, Amazon Video, Vudu, Google Play, and Disney Movies Anywhere on April 10, 2015.
Walt Disney Studios Home Entertainment reissued The Phantom Menace on Blu-ray, DVD, and digital download on September 22, 2019. Additionally, all six films were available for 4K HDR and Dolby Atmos streaming on Disney+ upon the service's launch on November 12, 2019. This version of the film was released by Disney on 4K Ultra HD Blu-ray box set on March 31, 2020.
3D re-release
On September 28, 2010, it was announced that all six films in the series would be stereo-converted to 3D. These would be re-released in episode order, beginning with The Phantom Menace, which was released to cinemas on February 10, 2012. Prime Focus Limited did the conversion under the supervision of ILM. However, the 3D re-releases of Episodes II–VI were postponed or canceled after Lucasfilm was bought by The Walt Disney Company, who decided to focus on the development of Star Wars: The Force Awakens.
Lucas stated the 3D re-release was "just a conversion" of the film's 2011 Blu-ray release and no additional changes were made. Only a change to Anakin's magnetic wand during the podrace sceneits tip was sharpened to more accurately fit the original 2D photography to the new 3D imagewas confirmed.
General Mills and Brisk were promotional partners in North America for the 3D re-release but promotion was limited. The film was extensively promoted in Japan; promotional products were sold by 7-Eleven, Domino's Pizza, Pepsi and Gari-Gari Kun. Kellogg's promoted the film internationally, and French restaurant Quick launched three Star Wars-themed burgers. Lucasfilm also partnered with Variety, the Children's Charity to raise funds for children through the sale of a special edition badge.
Reception
Critical response
Following an advance screening on Saturday, May 8, 1999, several newspapers broke an agreement with Fox and published reviews of the film on Sunday, May 9. In a front page review, the Los Angeles Daily News gave it 3½ stars calling it "pretty good" overall and "outstanding in many parts". The New York Daily News was less positive giving it 2½ stars. Variety also made its review by Todd McCarthy available on the Sunday with McCarthy calling it "the most widely anticipated and heavily hyped film of modern times" but said that the film "can scarcely help being a letdown on some levels, but it’s too bad that it disappoints on so many" and that "it is neither captivating nor transporting, for it lacks any emotional pull, as well as the sense of wonder and awe that marks the best works of sci-fi/fantasy".
On review aggregator Rotten Tomatoes, the film has an approval rating of 52% based on 233 reviews, with an average rating of 5.9/10. The site's critical consensus reads, "Burdened by exposition and populated with stock characters, The Phantom Menace gets the Star Wars prequels off to a bumpy – albeit visually dazzling – start." As of December 2021, the film is tied with The Rise of Skywalker as the lowest-rated live-action film of the Star Wars series. On Metacritic, the film has a weighted average score of 51 out of 100, based on 36 critics, indicating "mixed or average reviews". Audiences polled by CinemaScore gave the film an average grade of "A–" on an A+ to F scale.
Many aspects of the scripting and characters were criticized, especially that of Jar Jar Binks, who was regarded by many members of the older fan community as toyetica merchandising opportunity rather than a serious character. Kenneth Turan of the Los Angeles Times described Binks as "a major miscue, a comic-relief character who's frankly not funny". Drew Grant of Salon wrote, "Perhaps the absolute creative freedom director George Lucas enjoyed while dreaming up the flick's 'comic' reliefwith no studio execs and not many an independently minded actor involvedis a path to the dark side."
Conversely, Roger Ebert of the Chicago Sun-Times gave it three-and-a-half stars out of four and called it "an astonishing achievement in imaginative filmmaking" and said, "Lucas tells a good story." Ebert also wrote that, "If some of the characters are less than compelling, perhaps that's inevitable" because it is the opening film in the new trilogy. He concluded his review by saying that rather than Star Trek films, filmmakers could "[g]ive me transparent underwater cities and vast hollow senatorial spheres any day". Owen Gleiberman of Entertainment Weekly gave the film a "B" grade and complimented Liam Neeson's performance and the action scenes. In an Entertainment Weekly review for the DVD release, Marc Bernardin gave the film a "C–", calling it "haplessly plotted, horribly written, and juvenile". ReelViews' James Berardinelli wrote, "Looking at the big picture, in spite of all its flaws, The Phantom Menace is still among the best 'bang for a buck' fun that can be had in a movie theater," and said the film was a "distinct improvement" over Return of the Jedi.
Andrew Johnston of Time Out New York wrote, "Let's face it: no film could ever match the expectations some have for Episode IThe Phantom Menace. Which isn't to say it's a disappointment: on the contrary, it's awesomely entertaining, provided you accept it on its own terms ... Like the original film, it's a Boy's Own adventure yarn with a corny but irresistible spiritual subtext. The effects and production design are stunning, but they always serve the story, not the other way around." Susan Wloszczyna of USA Today said that the film did "plenty right" and praised the characters Darth Maul and Watto. David Cornelius of efilmcritic.com said that the film's better moments "don't merely balance out the weaker onesthey topple them". Colin Kennedy of Empire magazine said that despite problems with pacing and writing, "there is still much pleasure to be had watching our full-blown Jedi guides in action". He praised the visuals and Liam Neeson's performance, and said that the duel between Darth Maul and the Jedi was "the saga's very best lightsaber battle".Empire magazine ranked The Phantom Menace on its list of "500 Greatest Movies of All Time", while Entertainment Weekly and Comcast included the film on their lists of the worst movie sequels. James Berardinelli wrote, "The Phantom Menace was probably the most overhyped motion picture of the last decade (if not longer), and its reputation suffered as a result of its inability to satisfy unreasonable expectations." William Arnold of the Seattle Post-Intelligencer agreed that the film's massive hype caused many of the negative reactions, saying, "it built expectations that can't possibly be matched and scuttled [the] element of storytelling surprise". He also said that the film was "well made and entertaining" and was much better than similar box office fare released around that year, such as The Mummy and The Matrix. Ewan McGregor said in 2002 that he was "slightly disappointed" that the film was "kind of flat" and believed the next film in the franchise would have "much more humor and...color."
The introduction of midi-chlorians—microscopic organisms that mediate use of the Force—has been regarded as controversial. Some viewed it as a concept that negates the Force's spiritual quality. Film historian Daniel Dinello says, "Anathema to Star Wars fanatics who thought they reduced the Force to a kind of viral infection, midi-chlorians provide a biological interface, the link between physical bodies and spiritual energy." Religion expert John D. Caputo writes, "In the 'Gospel according to Lucas', a world is conjured up in which the intractable oppositions that have tormented religious thinkers for centuries are reconciled ... The gifts that the Jedi masters enjoy have a perfectly plausible scientific basis, even if its ways are mysterious: their bodily cells have a heavier than usual concentration of 'midi-chlorians'."
There has been some controversy over whether several alien characters reflect racial stereotypes. For example, the oafish, slow-witted Jar Jar Binks has long droopy ears reminiscent of dreadlocks and spoke with what many perceived as a Caribbean patois reminiscent of Jamaican Creole. Drew Grant describes the character as "[s]ervile and cowardly ... a black minstrel-ish stereotype on par with Stepin Fetchit." Georgetown University professor of African-American studies Michael Eric Dyson says that the entire Gungan species seems suggestive of a primitive African tribe, with Boss Nass portrayed as "a fat, bumbling ... caricature of a stereotypical African tribal chieftain." The greedy and corrupt Neimoidians of the Trade Federation have been noted as resembling East Asian stereotypes, and the unprincipled trader Watto has been interpreted as a Jewish stereotype reminiscent of Charles Dickens' character Fagin. Lucas has denied all of these implications, instead criticizing the American media for using opinions from the Internet as a reliable source for news stories. Lucas added that it reflects more the racism of the commenters than it does the movie; however, animator Rob Coleman said ahead of the film's release that Watto's mannerisms were inspired by footage of Alec Guinness as Fagin in Oliver Twist.
Box office
Despite its mixed critical reception, The Phantom Menace was a financial success, breaking many box office records in its debut, and beating out The Mummy by taking number 1. It broke The Lost World: Jurassic Parks records for the largest single-day gross for taking more than $28 million in the opening day and fastest to gross $100 million in five days. It grossed $64.8 million in its opening weekend, the second-ever highest at the time. It also became the quickest film to reach the $200 million and $300 million marks, surpassing Independence Day (1996) and Titanic (1997) respectively. The Phantom Menace was 1999's most successful film, staying in the Top 10 until August 5 (11 weeks total), earning $431.1 million in the United States and Canada. Box Office Mojo estimates that the film sold over 84.8 million tickets in the US in its initial theatrical run.
The film set an opening record in Japan, grossing $12.2 million in its first two days from 403 screens. In the UK, the film also set an opening record with £9.5 million in its opening weekend (including previews). It also grossed a record $11 million in its opening weekend in Germany. Outside the United States and Canada, the film grossed over $10 million in Australia ($25.9 million), Brazil ($10.4 million), France and Algeria ($43 million), Germany ($53.9 million), Italy ($12.9 million), Japan ($109.9 million), Mexico ($12 million), Spain ($25 million), and the United Kingdom and Ireland ($81.9 million). Its overseas total was $493.2 million, taking its worldwide total to $924.3 million. At that time, the film was the third highest-grossing film in North America behind Titanic and Star Wars (1977), and the second highest-grossing film worldwide behind Titanic without adjusting for inflation of ticket prices. When adjusted for ticket price inflation, it ranked as the 19th-highest-grossing film domestically, making it the fourth Star Wars film to be in the Inflation-Adjusted Top 20.
After its 3D re-release in 2012, the worldwide box office gross exceeded $1 billion. Although in the intervening years, the film had lost some of its rankings in the lists of highest-grossing films, the 3D re-release returned it to the worldwide all-time Top 10 for several months. In North America, its revenues overtook those of the original Star Wars as the saga's highest-grossing film when not adjusting for inflation of ticket prices, and is the tenth highest-grossing film in North America . In North America, its ranking on the Adjusted for Ticket Price Inflation list climbed to 16th placeone place behind Return of the Jedi. The 3D , which premiered in February 2012, earned $43 million$22.5 million of which was in North Americaworldwide. The 3D re-release earned US$ worldwideincluding $43.5 million in North Americaand has increased the film's overall box office takings to $474.5 million domestically, and $552.5 million in other territories. The film's earnings exceeded $1 billion worldwide on February 22, 2012, making it the first Star Wars film and the 11th film in historyexcluding inflationto do so.
AccoladesThe Phantom Menace was nominated for three Academy Awards: Best Sound Effects Editing, Best Visual Effects, and Best Sound (Gary Rydstrom, Tom Johnson, Shawn Murphy and John Midgley); all of which went to The Matrix. The film won Saturn Awards for Best Costumes and Best Special Effects, the MTV Movie Award for Best Action Scene, and a Young Artist Award for Jake Lloyd's performance. It was also nominated foramong othersthe BAFTAs for Visual Effects and Sound, and the Grammy Award for Best Score Soundtrack for Visual Media. The film did however receive seven Golden Raspberry Award (Razzie) nominations for Worst Picture, Worst Director, Worst Screenplay, Worst Supporting Actor (Jake Lloyd as Anakin), Worst Supporting Actress (Sofia Coppola as Saché), Worst Screen Couple (Jake Lloyd and Natalie Portman), and Jar Jar Binks actor Ahmed Best won the Worst Supporting Actor category.
The film was nominated Choice Drama Movie, Choice Movie Villain and Choice Summer Movie at 1999 Teen Choice Awards, but both lost to Cruel Intentions and Big Daddy, respectively.
Legacy
A month after the film's release, "Weird Al" Yankovic released the parody song and music video "The Saga Begins", in which he interprets the film's plot from Obi-Wan's point of view to the tune of "American Pie"; this was included as a bonus feature on a 2011 Star Wars Blu-ray. In a 2018 Saturday Night Live comedy rap video, Natalie Portman reprised her appearance as Queen Amidala from The Phantom Menace and defended the prequel trilogy.
Darth Maul's lightsaber-fighting style served as the inspiration for the 2003 viral video Star Wars Kid. Maul, who appears to die in The Phantom Menace, was resurrected for the animated series Star Wars: The Clone Wars, and also appears in Star Wars Rebels and Solo: A Star Wars Story. In 2012, IGN named Maul the 16th greatest Star Wars character. A similar weapon to his dual-bladed lightsaber appears in Star Wars: The Rise of Skywalker.
Sequels
A sequel, Attack of the Clones, was released in 2002. The story continues a decade later with Anakin grown to adulthood, played by Hayden Christensen. A second sequel, Revenge of the Sith, was released in 2005.
References
Footnotes
Citations
Sources
Further reading
Michael J. Hanson, Michael J. & Kay, Max S. (2000). Star Wars: The New Myth. Xlibris,
McDonald, Paul F. (2013). The Star Wars Heresies: Interpreting The Themes, Symbols and Philosophies of Episodes I-III. McFarland,
Bortolin, Matthew (2005). The Dharma of Star Wars''. Wisdom Publications,
External links
at StarWars.com
at Lucasfilm.com
Star Wars Episode I: The Phantom Menace at Corona's Coming Attractions
1990s science fiction action films
1990s science fiction adventure films
1999 films
1999 science fiction films
20th Century Fox films
3D re-releases
Alien invasions in films
American films
American science fiction war films
Films about children
Films about elections
Films about slavery
Films directed by George Lucas
Films featuring puppetry
Films produced by Rick McCallum
Films scored by John Williams
Films shot at Warner Bros. Studios, Leavesden
Films shot in Hertfordshire
Films shot in Italy
Films shot in Sydney
Films shot in Tunisia
Films with screenplays by George Lucas
Films with underwater settings
Films using motion capture
Golden Raspberry Award winning films
Lucasfilm films
Prequel films
Phantom Menace, The
Underwater civilizations in fiction |
null | null | Star Wars: Episode II – Attack of the Clones | eng_Latn | Star Wars: Episode II – Attack of the Clones is a 2002 American epic space opera film directed by George Lucas and written by Lucas and Jonathan Hales. The sequel to The Phantom Menace (1999), it is the fifth film in the Star Wars film series and second chronological chapter of the "Skywalker Saga". The film stars Ewan McGregor, Natalie Portman, Hayden Christensen, Ian McDiarmid, Samuel L. Jackson, Christopher Lee, Anthony Daniels, Kenny Baker, and Frank Oz.
The story is set ten years after The Phantom Menace, as thousands of planetary systems slowly secede from the Galactic Republic and join the newly-formed Confederacy of Independent Systems, led by former Jedi Master Count Dooku. With the galaxy on the brink of civil war, Obi-Wan Kenobi investigates a mysterious assassination attempt on Senator Padmé Amidala, which leads him to uncover a clone army in service of the Republic and the truth behind the Separatist movement. Meanwhile, his apprentice Anakin Skywalker is assigned to protect Amidala and develops a secret relationship with her. Soon, the trio witness the onset of a new threat to the galaxy: the Clone Wars.
Development of Attack of the Clones began in March 2000, some months after the release of The Phantom Menace. By June 2000, Lucas and Hales completed a draft of the script, and principal photography took place from June to September 2000. The film crew primarily shot at Fox Studios Australia in Sydney, Australia, with additional footage filmed in Tunisia, Spain, and Italy. It was one of the first motion pictures shot completely on a high-definition digital 24-frame system.
The film was released in the United States on May 16, 2002. It received mixed reviews, with some critics hailing it as an improvement over its predecessor The Phantom Menace and others considering it the worst installment of the franchise. The film was praised for an increased emphasis on action, visual effects, musical score, and costume design, but criticized for the screenplay, Christensen's performance, romantic scenes, and underdeveloped characters. It performed well in the box office, making over $645 million worldwide; however, it became the first Star Wars film to be outgrossed in its year of release, placing third domestically and fourth-highest-grossing worldwide. Revenge of the Sith (2005) followed Attack of the Clones, concluding the Star Wars prequel trilogy.
Plot
Ten years after the battle at Naboo, the Galactic Republic is threatened by a Separatist movement organized by former Jedi Master ,Count Dooku. Senator Padmé Amidala comes to Coruscant to vote on a motion to create an army to assist the Jedi against the threat. Narrowly avoiding an assassination attempt upon her arrival, she is placed under the protection of Jedi Master Obi-Wan Kenobi and his apprentice Anakin Skywalker. The pair thwart a second attempt on Padmé's life and subdue the assassin, Zam Wesell, who is killed by her employer, a bounty hunter, before she can reveal his identity. The Jedi Council instructs Obi-Wan to find the bounty hunter, while Anakin is tasked to protect Padmé and escort her back to Naboo, where the two fall in love in spite of the Jedi Code that forbids attachments.
Obi-Wan's search leads him to the mysterious ocean planet of Kamino, where he discovers an army of clones being produced for the Republic under the name of Sifo Dyas, a deceased Jedi Master, with bounty hunter Jango Fett serving as their genetic template. Obi-Wan meets with Jango, who reveals that the clones were the idea of a man called Tyranus. Obi-Wan deduces Jango to be the bounty hunter he is seeking, and after a brief battle, places a homing beacon on Jango's ship, Slave I. He then follows Jango and his clone son, Boba, to the planet Geonosis. Meanwhile, Anakin is troubled by visions of his mother, Shmi, in pain and decides to return to his homeworld of Tatooine with Padmé to save her. Watto reveals that he sold Shmi to moisture farmer Cliegg Lars, who then freed and married her. Cliegg tells Anakin that she was abducted by Tusken Raiders weeks earlier and is likely dead. Determined to find his mother, Anakin ventures out and finds her at the Tusken campsite, still barely alive. After she dies in his arms, an enraged Anakin massacres the tribe. He later declares to Padmé that he will find a way to prevent the deaths of those he loves.
On Geonosis, Obi-Wan discovers a Separatist gathering led by Count Dooku, who is developing a droid army with Trade Federation Viceroy Nute Gunray and ordered the attempts on Padmé's life. Obi-Wan transmits his findings to the Jedi Council but is captured by Separatist droids. Dooku meets Obi-Wan in his cell and explains his role in the Confederacy of Independent Systems' formation, while implying that the Sith Lord Darth Sidious is in control of a large portion of the Galactic Senate. He then invites Obi-Wan to join him and stop Sidious. When Obi-Wan refuses, Dooku claims that Obi-Wan's late master and Dooku's former apprentice Qui-Gon Jinn would have, had he been alive. Meanwhile, Senate Representative Jar Jar Binks proposes a successful vote to grant emergency powers to Chancellor Palpatine, allowing the clone army to be authorized.
Anakin and Padmé head to Geonosis to rescue Obi-Wan, but Anakin loses his lightsaber and they are captured by Jango. Dooku sentences the trio to death, but they are saved by a battalion of clone troopers led by Yoda, Mace Windu, and other Jedi. Windu beheads Jango during the ensuing battle. Obi-Wan and Anakin intercept Dooku, and they engage in a lightsaber duel. Dooku injures Obi-Wan and severs Anakin's right arm, but Yoda intercepts and defends them. Dooku uses the Force to distract Yoda and escapes to Coruscant, where he delivers the plans of a super-weapon to Sidious, who addresses Dooku by his Sith name Tyranus. The Council is left disturbed by Dooku's claim of Sidious controlling the Senate. As the Jedi acknowledge the beginning of the Clone Wars, Anakin is fitted with a robotic hand and marries Padmé on Naboo, with and R2-D2 as the only witnesses.
Cast
Ewan McGregor as Obi-Wan Kenobi: A Jedi Knight and mentor to his Padawan learner, Anakin Skywalker, who investigates the assassination attempt of Padmé, leading him to discover the production of a Clone Army for the Galactic Republic. In the 10 years since The Phantom Menace, he has grown wiser and more powerful in the use of the Force.
Natalie Portman as Senator Padmé Amidala: Former Queen of Naboo, who has recently been elected the planet's Senator, and Anakin's love interest.
Hayden Christensen as Anakin Skywalker: A former slave from Tatooine and Obi-Wan's gifted Padawan apprentice who is assigned to protect Padmé with whom he falls in love. He is believed to be the "chosen one" of Jedi prophecy destined "to bring balance to The Force." In the 10 years since The Phantom Menace, he has grown powerful but arrogant, and believes that Obi-Wan is holding him back. A large search for an actor to portray Anakin Skywalker was performed. Lucas auditioned various actors, mostly unknown, before casting Christensen. Among the many established actors who auditioned were Jonathan Brandis, Devon Sawa, Topher Grace, Joshua Jackson, Ryan Phillippe, Colin Hanks, and Paul Walker. Leonardo DiCaprio also met with Lucas for the role, but was "definitely unavailable" according to DiCaprio publicist Ken Sunshine. Co-star Natalie Portman later told Time magazine that Christensen "gave a great reading. He could simultaneously be scary and really young."
Christopher Lee as Count Dooku / Darth Tyranus: A former Jedi Master and the old mentor of Obi-Wan's late master Qui-Gon Jinn, who is now the puppet leader of the Separatist movement as well as Darth Sidious' new Sith apprentice and a suspect in Obi-Wan's investigation.
Samuel L. Jackson as Mace Windu: A Jedi Master sitting on the Jedi Council who warily watches the Galactic Senate's politics.
Frank Oz as the voice of Yoda: The centuries-old Jedi Grandmaster of an unknown alien species. In addition to leading the Jedi Council, Yoda is the instructor for the young Jedi Padawans.
Ian McDiarmid as Supreme Chancellor Palpatine / Darth Sidious: A former senator from Naboo, as well as a secret Sith Lord, who amasses vast emergency powers as the Chancellor of the Galactic Republic upon the outbreak of the Clone Wars.
Pernilla August as Shmi Skywalker: Anakin's mother and a former slave.
Temuera Morrison as Jango Fett: A bounty hunter who gave his DNA for use by the cloning facilities on Kamino for the creation of the clone army. In addition to his wage, he requested an unaltered clone for himself to raise as his son – Boba Fett.
Anthony Daniels as C-3PO: A protocol droid built by Anakin as a child who now serves the Lars family. He later leaves Tatooine to accompany Anakin and Padmé.
Daniels also makes a cameo appearance as Dannl Faytonni: An attendee of the Coruscant Outlander Club who witnesses Anakin and Obi-Wan capturing Zam Wesell
Kenny Baker as R2-D2: Anakin's astromech droid who often accompanies him and Obi-Wan on missions.
Daniel Logan as Boba Fett: Jango Fett's clone and adopted son.
Leeanna Walsman as Zam Wesell: A bounty hunter and partner of Jango Fett, who was given the task of assassinating Padmé. Although her appearance was human, she was actually a Clawdite, who could shapeshift.
Silas Carson as Nute Gunray and Ki-Adi-Mundi: Gunray is the Nemoidian Viceroy of the Trade Federation and one of the leaders of the Separatist movement, who puts a hit on Padmé as revenge for his loss against her people on Naboo in The Phantom Menace. Ki-Adi-Mundi is a Cerean Jedi Master sitting on the Jedi Council.
Ahmed Best as Delegate Jar Jar Binks: Padmé's loyal Gungan friend whom she appoints Representative of Naboo in the Galactic Senate.
Best also makes a cameo appearance as Achk Med-Beq: An attendee of the Coruscant Outlander Club who witnesses Anakin and Obi-Wan capturing Zam Wesell
Jimmy Smits as Bail Organa: A senator from Alderaan.
Jay Laga'aia as Gregar Typho: Padmé's newly appointed captain of security.
Oliver Ford Davies as Sio Bibble: Governor of Naboo.
Rose Byrne as Dormé: Handmaiden of Padmé.
Jack Thompson as Cliegg Lars: Shmi's husband, Owen's father and Anakin's stepfather.
Joel Edgerton as Owen Lars: Cliegg's son, Shmi’s stepson, and Anakin's stepbrother.
Bonnie Piesse as Beru Whitesun: Owen's girlfriend.
Andy Secombe as Watto: Anakin and Shmi's former slave owner.
Ronald Falk as Dexter Jettster: Obi-Wan's Besalisk friend who runs a diner on Coruscant and informs him about Kamino.
Alethea McGrath as Jocasta Nu: The librarian at the Jedi Temple who helps Obi-Wan gain information about Kamino.
Liam Neeson as Qui-Gon Jinn (statue and voice cameo only)
E! reported that Lucas had asked NSYNC to film a small background cameo appearance, in order to satisfy his daughters. They were subsequently cut out of the film in post-production, although briefly visible during a crowd shot from above. The end credits erroneously list Alan Ruscoe as playing Neimoidian senator Lott Dod. The character was actually another Neimoidian, played by an uncredited David Healey and voiced by Christopher Truswell. Liam Neeson reprised his role as Qui-Gon Jinn from The Phantom Menace, who only appears as a disembodied ghostly voice heard by Anakin through the Force as he was slaughtering the Tusken Raiders; Qui-Gon Jinn also appears in the form of a statue in his likeness within the Jedi Archives during one scene. Fiona Johnson reprised her The Matrix (1999) role as the Woman in the Red Dress in an Easter egg cameo appearance, with her character named "Hayde Gofai" in later Star Wars media, briefly offering Anakin a seductive look at the Outlander Club.
Production
Writing
After the mixed critical response to The Phantom Menace, Lucas was hesitant to return to the writing desk. In March 2000, just three months before the start of principal photography, Lucas finally completed his rough draft for Episode II. Lucas continued to iterate on his rough draft, producing a proper first and second draft. For help with the third draft, which would later become the shooting script, Lucas brought on Jonathan Hales, who had written several episodes of The Young Indiana Jones Chronicles for him, but had limited experience writing theatrical films. The final script was completed just one week before the start of principal photography.
As an in-joke, the film's working title was Jar Jar's Great Adventure, a sarcastic reference to the negative fan response to the Episode I character.
In writing The Empire Strikes Back, Lucas initially decided that Lando Calrissian was a clone and came from a planet of clones which caused the "Clone Wars" mentioned by Obi-Wan Kenobi in A New Hope; he later came up with an alternate concept of an army of clone shocktroopers from a remote planet which were used by the Republic as an army in the war that followed.
Filming
Principal photography occurred between June 26, 2000 and September 20, 2000 at Fox Studios Australia in Sydney. Location shooting took place in the Tunisian desert, at the Plaza de España in Seville, London, China, Vancouver, San Diego, and Italy (Villa del Balbianello on Lake Como, and in the former royal Palace of Caserta). At his own personal request, Samuel L. Jackson's character Mace Windu received a lightsaber that emits a purple glow, as opposed to traditional blue and green for "good guys" and red for "bad guys". Reshoots were performed in March 2001. During this time, a new action sequence was developed featuring the droid factory after Lucas had decided that the film lacked a quick enough pace in the corresponding time-frame. The sequence's previsualization was rushed, and the live-action footage was shot within four and a half hours. Because of Lucas' method of creating shots through various departments and sources that are sometimes miles and years apart from each other, Attack of the Clones became the first film ever to be produced through what Rick McCallum called "virtual filmmaking".
Like The Phantom Menace, Attack of the Clones furthered technological development, effectively moving Hollywood into the "digital age" with the use of the HDW-F900, developed by Sony and Panavision, a digital camera using an HD digital 24-frame system. This spawned controversy over the benefits and disadvantages of digital cinematography that continues as more filmmakers "convert" to digital filmmaking while many filmmakers oppose it. In contrast to previous installments, for which scenes were shot in the Tunisian desert in temperatures up to 125 °F (51 °C), the camera would still run without complications. Lucas had stated that he wished to film The Phantom Menace on this format but Sony was unable to build the cameras quickly enough. In 2002, Attack of the Clones became the third film to be released that was shot entirely on a 24p digital camera (preceded by 2001's Jackpot and Vidocq). The cameras record in the 16:9 HDCAM format (1080p), although the image was cropped to a 2.40:1 widescreen ratio. The area above and below the 2.40 extraction area was available for Lucas to reframe the picture as necessary in post-production.
Despite Lucas' efforts to persuade movie theaters to switch to digital projectors for viewing of Episode II, few theaters did.
Visual effects
The film relied almost solely on digital animatics as opposed to storyboards in order to previsualize sequences for editing early on in the film's production. While Lucas had used other ways of producing motion-based storyboards in the past, after The Phantom Menace the decision was made to take advantage of the growing digital technology. The process began with Ben Burtt's creation of what the department dubbed as "videomatics", so called because they were shot on a household videocamera. In these videomatics, production assistants and relatives of the department workers acted out scenes in front of greenscreen. Using computer-generated imagery (CGI), the previsualization department later filled in the green screen with rough background footage. Burtt then cut together this footage and sent it off to Lucas for changes and approval. The result was a rough example of what the final product was intended to be. The previsualization department then created a finer version of the videomatic by creating an animatic, in which the videomatic actors, props, and sets were replaced by digital counterparts to give a more precise, but still rough, look at what would eventually be seen. The animatic was later brought on set and shown to the actors so that they could understand the concept of the scene they were filming in the midst of the large amount of bluescreen used. Unlike most of the action sequences, the Battle of Geonosis was not story-boarded or created through videomatics but was sent straight to animatics after the department received a small vague page on the sequence. The intent was to create a number of small events that would be edited together for pacing inside the finished film. The animatics department was given a free hand regarding events to be created within the animatic; Lucas only asked for good action shots that he could choose from and approve later.
In addition to introducing the digital camera, Attack of the Clones emphasized "digital doubles" as computer-generated models that doubled for actors, in the same way that traditional stunt doubles did. It also furthered the authenticity of computer-generated characters by introducing a new, completely CGI-created version of the character Yoda. Rob Coleman and John Knoll prepared two tests featuring a CGI-animated Yoda using audio from The Empire Strikes Back. Yoda's appearance in Episode V also served as the reference point for the creation of the CGI Yoda; Lucas repeatedly stated to the animation department that "the trick" to the animation of the CGI Yoda was to make him like the puppet from which he was based, in order to maintain a flow of continuity. Frank Oz (voice and puppeteer for Yoda in the original trilogy and The Phantom Menace) was consulted; his main piece of advice was that Yoda should look extremely old, sore, and frigid. Coleman later explained the process of making the digital Yoda like the puppet version, by saying "When Frank [Oz] would move the head, the ears would jiggle. If we hadn't put that in, it wouldn't look like Yoda." Because of the acrobatics of the lightsaber fight between Count Dooku and Yoda, the then 78-year-old Christopher Lee relied on a stunt double to perform the most demanding scenes instead. Lee's face was superimposed onto the double's body in all shots other than close-ups, which he performed himself. Lucas often called the duel crucial to the animation department, as it had such potential to be humorous rather than dramatic.
Music
The soundtrack to the film was released on April 23, 2002 by Sony Classical Records. The music was composed and conducted by John Williams, and performed by the London Voices and London Symphony Orchestra. The soundtrack recreates "The Imperial March" from the film The Empire Strikes Back for its first chronological appearance in Attack of the Clones, even though a hint of it appeared in the previous movie in one of the final scenes. A music video for the main theme "Across the Stars" was produced specifically for the DVD.
On March 15, 2016, a limited edition vinyl version of the soundtrack was released. Only 1,000 copies were pressed initially.
Themes
Lucas has noted that Palpatine's rise to power is very similar to that of Adolf Hitler in Nazi Germany; as Chancellor of Germany, the latter was granted emergency powers, as is Palpatine. Comparisons have been made to Octavian—who became Augustus, the first emperor of Rome—and to Napoleon Bonaparte, who rose to power in France from 1796 to 1799. Octavian was responsible for the deaths of hundreds of political opponents well before he was granted tribunician powers; Bonaparte was appointed First Consul for life (and later Emperor) by the French Consulate after a failed attempt on his life and the subsequent coup of 18 Brumaire in 1799. References to the American Civil War can also be discerned.
War journalism, combat films, and footage of World War II combat influenced the documentary-style camera work of the Battle of Geonosis, even to the point that hand-held shakes were digitally added to computer-generated sequences.
English scholar Anne Lancashire describes Attack of the Clones as "thoroughly political in its narrative", to the point that interpersonal relations are made subordinate to the political drama that unfolds, and "a critique of the increasing role played by economic and political appetite in contemporary First World international politics in general". In this political drama, the Trade Federation, the former idealist Dooku, and Palpatine "[represent] the economic and political greed and ambition ... of the political and business classes", while the intuition of the Jedi has been clouded by the dark side of the Force. The cityscape of Coruscant, the location of the Jedi Temple, is a dystopian environment that refers to 1982's Blade Runner. Nevertheless, the Jedi endure as the heroes; Obi-Wan's role has been noted as similar to that of James Bond, and Zam Wesell's attempt on Padmé's life is similar to a scene in the first 007 film, Dr. No.
The prequel trilogy films often refer to the original trilogy in order to help connect the films together. Lucas has often referred to the films as a long poem that rhymes. Such examples include the line "I have a bad feeling about this", a phrase used in each film, and lightsaber duels which almost always occur over a pit. As with Attack of the Clones, The Empire Strikes Back was the middle film in a trilogy, and of the original trilogy films, The Empire Strikes Back is the object of the most references in Attack of the Clones. In both films, an asteroid field is the backdrop of a major star battle in the middle of the film. Obi-Wan escapes Jango Fett by attaching his spacecraft to an asteroid in order to disappear from the enemy sensors; Han Solo uses a similar tactic by attaching the Millennium Falcon to a Star Destroyer in The Empire Strikes Back. As a retcon, John Knoll confirms on the film's DVD commentary that Boba Fett, who would later catch Solo in the act in The Empire Strikes Back, "learned his lesson" from the events of Attack of the Clones. The Galactic Republic's clone troopers also establish the origin of the stormtroopers that play an important role in the original trilogy.
Release
After a teaser trailer premiered with the film Monsters, Inc., a new trailer for the film aired on the Fox Network on March 10, 2002 between Malcolm in the Middle and The X-Files, then on March 15, 2002 with Ice Age in theatres and was made available on the official Star Wars web site the same day. The outplacement firm Challenger, Gray & Christmas from Chicago predicted before the film's release that U.S. companies could lose more than $319 million in productivity due to employees calling in sick and then heading to theaters to see the film.
The film premiered as part of the inaugural Tribeca Film Festival at the BMCC Performing Arts Center, 199 Chambers St. in New York City at a Sunday, May 12 set of screenings benefitting the Children's Aid Society, a charity supported by George Lucas. Attack of the Clones was then screened out of competition at the 2002 Cannes Film Festival, before getting a worldwide theatrical release on May 16, 2002. The film was also later released in IMAX theaters; the film had not been filmed for IMAX but was "up converted" with the digital remastering process. Because of the technical limitations of the IMAX projector at the time, an edited, 120-minute version of the film was presented.
Before the film's release, there was a string of controversies regarding copyright infringement. In 2000, an underground organization calling itself the Atlas Group, based in Perth, Western Australia offered a copy of the screenplay, with an asking price of US$100,000, to various fan sites and media organizations, including TheForce.Net. The scheme was subsequently reported to Lucasfilm Ltd. by the fan site.
An unauthorized copy was allegedly made at a private showing, using a digital recorder that was pointed at the screen. This copy spread over the internet, and analysts predicted up to a million fans would have seen the film before the day of its release. In addition, authorities seized thousands of bootlegs throughout Kuala Lumpur before the film opened.
Home media
Attack of the Clones was released on DVD and VHS on November 12, 2002. On the first day of release, over 4 million DVD copies were sold, becoming the third highest single-day DVD sales behind Monsters, Inc. and Spider-Man. This THX certified two-disc DVD release consists of widescreen and pan and scan fullscreen versions. The set contains one disc with the film and the other one with bonus features. The first disc features three randomized selected menus, which are Coruscant, Kamino and Geonosis. There is an Easter egg located in the options menu. When the THX Optimizer is highlighted, the viewer can press 1-1-3-8. By doing this, some bloopers and DVD credits will be shown. The DVD also features an audio commentary from director George Lucas, producer Rick McCallum, editor and sound designer Ben Burtt, ILM animation director Rob Coleman, and ILM visual effects supervisors Pablo Helman, John Knoll, and Ben Snow. Eight deleted scenes are included along with multiple documentaries, which include a full-length documentary about the creation of digital characters and two others that focus on sound design and the animatics team. Three featurettes examine the storyline, action scenes, and love story, and a set of 12 short web documentaries cover the overall production of the film.
The Attack of the Clones DVD also features a trailer for a mockumentary-style short film known as R2-D2: Beneath the Dome. Some stores offered the full mockumentary as an exclusive bonus disc for a small extra charge. The film gives an alternate look at the "life" of the droid R2-D2. The story, which Lucas approved, was meant to be humorous.
The film was re-released in a prequel trilogy DVD box set on November 4, 2008.
The six-film Star Wars saga was released on Blu-ray Disc on September 16, 2011 in three different editions.
On April 7, 2015, Walt Disney Studios, 20th Century Fox, and Lucasfilm jointly announced the digital releases of the six released Star Wars films. Attack of the Clones was released through the iTunes Store, Amazon Video, Vudu, Google Play, and Disney Movies Anywhere on April 10, 2015.
Walt Disney Studios Home Entertainment reissued Attack of the Clones on Blu-ray, DVD, and digital download on September 22, 2019. Additionally, all six films were available for 4K HDR and Dolby Atmos streaming on Disney+ upon the service's launch on November 12, 2019. This version of the film was released by Disney on 4K Ultra HD Blu-ray on March 31, 2020 whilst being re-released on Blu-ray and DVD.
3D re-release
On , it was announced that all six films in the series were to be stereo-converted to 3D, and re-released in chronological order beginning with The Phantom Menace which was released on . Attack of the Clones was originally scheduled to be re-released in 3D on , but was postponed due to Lucasfilm's desire to focus on Star Wars: The Force Awakens. However, the 3D presentation of the film was first shown at Celebration Europe II from July 26 to 28, 2013.
Reception
Critical response
On review aggregator Rotten Tomatoes, the film holds an approval rating of 65% based on 253 reviews, with an average rating of 6.59/10. The site's critical consensus reads, "Star Wars Episode II: Attack of the Clones benefits from an increased emphasis on thrilling action, although they're once again undercut by ponderous plot points and underdeveloped characters." On Metacritic, the film has a weighted average score of 54 out of 100, based on 39 critics, which indicates "mixed or average reviews". Audiences polled by CinemaScore gave the film an average grade of "A–" on an A+ to F scale, the same score as the previous film.
Numerous critics characterized the dialogue as "stiff" and "flat". The acting was also disparaged by some critics. Conversely, other critics felt fans would be pleased to see that Jar Jar Binks has only a minor role. Additionally, Jar Jar's attempts at comic relief seen in The Phantom Menace were toned down; instead, C-3PO reprised some of his bumbling traditions in that role. McGregor referred to the swordplay in the film as "unsatisfactory" when comparing it to the climactic duel in Revenge of the Sith as it neared release. ReelViews.net's James Berardinelli gave a positive review, saying "in a time when, more often than not, sequels disappoint, it's refreshing to uncover something this high-profile that fulfills the promise of its name and adds another title to a storied legacy."
Roger Ebert, who had praised the previous Star Wars films, gave Episode II only two out of four stars, noting "[As] someone who admired the freshness and energy of the earlier films, I was amazed, at the end of Episode II, to realize that I had not heard one line of quotable, memorable dialogue." About Anakin and Padme's relationship, Ebert stated, "There is not a romantic word they exchange that has not long since been reduced to cliché." Leonard Maltin, who also liked all of the previous installments, awarded two stars out of four to this endeavor as well, as seen in his Movie and Video Guide from the 2002 edition onward. Maltin cited an "overlong story" as reason for his dissatisfaction and added "Wooden characterizations and dialogue don't help." While TheWrap considered Attack of the Clones a "marginal improvement" over The Phantom Menace, Screen Rant ranked it as inferior to Episode I and as the worst Star Wars film overall, citing its romance story, characterization, and acting as weak points. Some film critics and fans criticized the CGI model of Yoda and how he has a lightsaber fight, with many loving it, while others believed it goes against his character.
Following suit with the series' previous installments, the Academy Awards nominated Attack of the Clones' Rob Coleman, Pablo Helman, John Knoll, and Ben Snow for Best Visual Effects at the 2003 Academy Awards, but the award ultimately went to The Lord of the Rings: The Two Towers. Natalie Portman was also honored at the Teen Choice Awards, and the film received an award for Best Fight at the MTV Movie Awards. In contrast, the film also received seven nominations from the Golden Raspberry Awards for Worst Picture, Worst Director (George Lucas), Worst Screenplay (George Lucas), Worst Supporting Actor (Hayden Christensen), Worst Supporting Actress (Natalie Portman), Worst Screen Couple (Hayden Christensen and Natalie Portman) and Worst Remake or Sequel. It took home two awards for Worst Screenplay (George Lucas) and Worst Supporting Actor (Hayden Christensen).
Box office
During its opening day, Attack of the Clones made $30.1 million worldwide. It would go on to make $116.3 million worldwide within four days. At that point, it was the second-fastest film to approach the $100 million mark. Plus, it had grossed over $80 million worldwide over the weekend. Attack of the Clones had the highest opening weekend for a 20th Century Fox film until 2003 when it was taken by X2. The film grossed $310,676,740 in North America and $338,721,588 overseas for a worldwide total of $649,398,328. Though a box office success, it was nevertheless overshadowed by the even greater box-office success of The Phantom Menace three years earlier. It was not the top-grossing film of the year, either in North America (where it finished in third place) or worldwide (where it was fourth), the first time that a Star Wars film did not have this distinction. In North America it was outgrossed by Spider-Man and The Lord of the Rings: The Two Towers, both of which were more favorably received by critics. Worldwide, it was also outgrossed by Harry Potter and the Chamber of Secrets. Adjusted for inflation, Attack of the Clones is the lowest-performing live-action Star Wars film at the North American box office, though is still among the 100 highest-grossing films of all time when adjusted for inflation. It sold an estimated 52,012,300 tickets in the US in its initial theatrical run.
Accolades
Novelizations
Two novels based on the movie were published, a tie-in junior novel by Scholastic, and a novelization written by R. A. Salvatore, which includes some unique scenes. A four-issue comic book adaptation was written by Henry Gilroy and published by Dark Horse Comics.
Sequel
A sequel titled Revenge of the Sith was released May 19, 2005 and was written and directed by George Lucas. It concludes the story of Anakin's transformation into Darth Vader following the death of Amidala and the destruction of all the Jedi except for Obi-Wan and Yoda.
References
Footnotes
Citations
Sources
External links
at
at
Star Wars Episode II: Attack of the Clones Official Production Notes
2000s adventure films
2000s English-language films
2000s science fiction adventure films
2002 films
2002 science fiction action films
2002 science fiction films
20th Century Fox films
American films
American science fiction war films
American sequel films
Films about cloning
Films about genetic engineering
Films directed by George Lucas
Films produced by Rick McCallum
Films scored by John Williams
Films shot at Elstree Film Studios
Films shot in Italy
Films shot in Spain
Films shot in Sydney
Films shot in Tunisia
Films with screenplays by George Lucas
Golden Raspberry Award winning films
IMAX films
Interquel films
Lucasfilm films
Prequel films
Attack of the Clones |
null | null | Sulfur dioxide | eng_Latn | Sulfur dioxide (IUPAC-recommended spelling) or sulphur dioxide (traditional Commonwealth English) is the chemical compound with the formula . It is a toxic gas responsible for the smell of burnt matches. It is released naturally by volcanic activity and is produced as a by-product of copper extraction and the burning of sulfur-bearing fossil fuels. Sulfur dioxide has a pungent smell like nitric acid.
Structure and bonding
SO2 is a bent molecule with C2v symmetry point group.
A valence bond theory approach considering just s and p orbitals would describe the bonding in terms of resonance between two resonance structures.
The sulfur–oxygen bond has a bond order of 1.5. There is support for this simple approach that does not invoke d orbital participation.
In terms of electron-counting formalism, the sulfur atom has an oxidation state of +4 and a formal charge of +1.
Occurrence
Sulfur dioxide is found on Earth and exists in very small concentrations and in the atmosphere at about 1 ppm.
On other planets, sulfur dioxide can be found in various concentrations, the most significant being the atmosphere of Venus, where it is the third-most abundant atmospheric gas at 150 ppm. There, it reacts with water to form clouds of sulfuric acid, and is a key component of the planet's global atmospheric sulfur cycle and contributes to global warming. It has been implicated as a key agent in the warming of early Mars, with estimates of concentrations in the lower atmosphere as high as 100 ppm, though it only exists in trace amounts. On both Venus and Mars, as on Earth, its primary source is thought to be volcanic. The atmosphere of Io, a natural satellite of Jupiter, is 90% sulfur dioxide and trace amounts are thought to also exist in the atmosphere of Jupiter.
As an ice, it is thought to exist in abundance on the Galilean moons—as subliming ice or frost on the trailing hemisphere of Io, and in the crust and mantle of Europa, Ganymede, and Callisto, possibly also in liquid form and readily reacting with water.
Production
Sulfur dioxide is primarily produced for sulfuric acid manufacture (see contact process). In the United States in 1979, 23.6 million metric tons (26,014,547 US short tons) of sulfur dioxide were used in this way, compared with 150 thousand metric tons (165,347 US short tons) used for other purposes. Most sulfur dioxide is produced by the combustion of elemental sulfur. Some sulfur dioxide is also produced by roasting pyrite and other sulfide ores in air.
Combustion routes
Sulfur dioxide is the product of the burning of sulfur or of burning materials that contain sulfur:
S + O2 → SO2, ΔH = −297 kJ/mol
To aid combustion, liquified sulfur (140–150 °C, 284-302 °F) is sprayed through an atomizing nozzle to generate fine drops of sulfur with a large surface area. The reaction is exothermic, and the combustion produces temperatures of 1000–1600 °C (1832–2912 °F). The significant amount of heat produced is recovered by steam generation that can subsequently be converted to electricity.
The combustion of hydrogen sulfide and organosulfur compounds proceeds similarly. For example:
2 H2S + 3 O2 → 2 H2O + 2 SO2
The roasting of sulfide ores such as pyrite, sphalerite, and cinnabar (mercury sulfide) also releases SO2:
4 FeS2 + 11 O2 → 2 Fe2O3 + 8 SO2
2 ZnS + 3 O2 → 2 ZnO + 2 SO2
HgS + O2 → Hg + SO2
4 FeS + 7O2 → 2 Fe2O3 + 4 SO2
A combination of these reactions is responsible for the largest source of sulfur dioxide, volcanic eruptions. These events can release millions of tons of SO2.
Reduction of higher oxides
Sulfur dioxide can also be a byproduct in the manufacture of calcium silicate cement; CaSO4 is heated with coke and sand in this process:
2 CaSO4 + 2 SiO2 + C → 2 CaSiO3 + 2 SO2 + CO2
Until the 1970s, commercial quantities of sulfuric acid and cement were produced by this process in Whitehaven, England. Upon being mixed with shale or marl, and roasted, the sulfate liberated sulfur dioxide gas, used in sulfuric acid production, the reaction also produced calcium silicate, a precursor in cement production.
On a laboratory scale, the action of hot concentrated sulfuric acid on copper turnings produces sulfur dioxide.
Cu + 2 H2SO4 → CuSO4 + SO2 + 2 H2O
From sulfites
Sulfites results by the action of aqueous base on sulfur dioxide:
SO2 + 2 NaOH → Na2SO3 + H2O
The reverse reaction occurs upon acidification:
H+ + HSO3− → SO2 + H2O
Reactions
Featuring sulfur in the +4 oxidation state, sulfur dioxide is a reducing agent. It is oxidized by halogens to give the sulfuryl halides, such as sulfuryl chloride:
SO2 + Cl2 → SO2Cl2
Sulfur dioxide is the oxidising agent in the Claus process, which is conducted on a large scale in oil refineries. Here, sulfur dioxide is reduced by hydrogen sulfide to give elemental sulfur:
SO2 + 2 H2S → 3 S + 2 H2O
The sequential oxidation of sulfur dioxide followed by its hydration is used in the production of sulfuric acid.
2 SO2 + 2 H2O + O2 → 2 H2SO4
Laboratory reactions
Sulfur dioxide is one of the few common acidic yet reducing gases. It turns moist litmus pink (being acidic), then white (due to its bleaching effect). It may be identified by bubbling it through a dichromate solution, turning the solution from orange to green (Cr3+ (aq)). It can also reduce ferric ions to ferrous.
Sulfur dioxide can react with certain 1,3-dienes in a cheletropic reaction to form cyclic sulfones. This reaction is exploited on an industrial scale for the synthesis of sulfolane, which is an important solvent in the petrochemical industry.
Sulfur dioxide can bind to metal ions as a ligand to form metal sulfur dioxide complexes, typically where the transition metal is in oxidation state 0 or +1. Many different bonding modes (geometries) are recognized, but in most cases, the ligand is monodentate, attached to the metal through sulfur, which can be either planar and pyramidal η1. As a η1-SO2 (S-bonded planar) ligand sulfur dioxide functions as a Lewis base using the lone pair on S. SO2 functions as a Lewis acids in its η1-SO2 (S-bonded pyramidal) bonding mode with metals and in its 1:1 adducts with Lewis bases such as dimethylacetamide and trimethyl amine. When bonding to Lewis bases the acid parameters of SO2 are EA = 0.51 and EA = 1.56.
Uses
The overarching, dominant use of sulfur dioxide is in the production of sulfuric acid.
Precursor to sulfuric acid
Sulfur dioxide is an intermediate in the production of sulfuric acid, being converted to sulfur trioxide, and then to oleum, which is made into sulfuric acid. Sulfur dioxide for this purpose is made when sulfur combines with oxygen. The method of converting sulfur dioxide to sulfuric acid is called the contact process. Several billion kilograms are produced annually for this purpose.
As a preservative
Sulfur dioxide is sometimes used as a preservative for dried apricots, dried figs, and other dried fruits, owing to its antimicrobial properties and ability to prevent oxidation, and is called E220 when used in this way in Europe. As a preservative, it maintains the colorful appearance of the fruit and prevents rotting. It is also added to sulfured molasses.
Sulfur dioxide was first used in winemaking by the Romans, when they discovered that burning sulfur candles inside empty wine vessels keeps them fresh and free from vinegar smell.
It is still an important compound in winemaking, and is measured in parts per million (ppm) in wine. It is present even in so-called unsulfurated wine at concentrations of up to 10 mg/L. It serves as an antibiotic and antioxidant, protecting wine from spoilage by bacteria and oxidation - a phenomenon that leads to the browning of the wine and a loss of cultivar specific flavors. Its antimicrobial action also helps minimize volatile acidity. Wines containing sulfur dioxide are typically labeled with "containing sulfites".
Sulfur dioxide exists in wine in free and bound forms, and the combinations are referred to as total SO2. Binding, for instance to the carbonyl group of acetaldehyde, varies with the wine in question. The free form exists in equilibrium between molecular SO2 (as a dissolved gas) and bisulfite ion, which is in turn in equilibrium with sulfite ion. These equilibria depend on the pH of the wine. Lower pH shifts the equilibrium towards molecular (gaseous) SO2, which is the active form, while at higher pH more SO2 is found in the inactive sulfite and bisulfite forms. The molecular SO2 is active as an antimicrobial and antioxidant, and this is also the form which may be perceived as a pungent odor at high levels. Wines with total SO2 concentrations below 10 ppm do not require "contains sulfites" on the label by US and EU laws. The upper limit of total SO2 allowed in wine in the US is 350 ppm; in the EU it is 160 ppm for red wines and 210 ppm for white and rosé wines. In low concentrations, SO2 is mostly undetectable in wine, but at free SO2 concentrations over 50 ppm, SO2 becomes evident in the smell and taste of wine.
SO2 is also a very important compound in winery sanitation. Wineries and equipment must be kept clean, and because bleach cannot be used in a winery due to the risk of cork taint, a mixture of SO2, water, and citric acid is commonly used to clean and sanitize equipment. Ozone (O3) is now used extensively for sanitizing in wineries due to its efficacy, and because it does not affect the wine or most equipment.
As a reducing agent
Sulfur dioxide is also a good reductant. In the presence of water, sulfur dioxide is able to decolorize substances. Specifically, it is a useful reducing bleach for papers and delicate materials such as clothes. This bleaching effect normally does not last very long. Oxygen in the atmosphere reoxidizes the reduced dyes, restoring the color. In municipal wastewater treatment, sulfur dioxide is used to treat chlorinated wastewater prior to release. Sulfur dioxide reduces free and combined chlorine to chloride.
Sulfur dioxide is fairly soluble in water, and by both IR and Raman spectroscopy; the hypothetical sulfurous acid, H2SO3, is not present to any extent. However, such solutions do show spectra of the hydrogen sulfite ion, HSO3−, by reaction with water, and it is in fact the actual reducing agent present:
SO2 + H2O ⇌ HSO3− + H+
As a fumigant
In the beginning of the 20th century, sulfur dioxide was used in Buenos Aires as a fumigant to kill rats that carried the Yersinia pestis bacterium, which causes bubonic plague. The application was successful, and the application of this method was extended to other areas in South America. In Buenos Aires, where these apparatuses were known as Sulfurozador, but later also in Rio de Janeiro, New Orleans and San Francisco, the sulfur dioxide treatment machines were brought into the streets to enable extensive disinfection campaigns, with effective results.
Biochemical and biomedical roles
Sulfur dioxide or its conjugate base bisulfite is produced biologically as an intermediate in both sulfate-reducing organisms and in sulfur-oxidizing bacteria, as well. The role of sulfur dioxide in mammalian biology is not yet well understood. Sulfur dioxide blocks nerve signals from the pulmonary stretch receptors and abolishes the Hering–Breuer inflation reflex.
It is considered that endogenous sulfur dioxide plays a significant physiological role in regulating cardiac and blood vessel function, and aberrant or deficient sulfur dioxide metabolism can contribute to several different cardiovascular diseases, such as arterial hypertension, atherosclerosis, pulmonary arterial hypertension, and stenocardia.
It was shown that in children with pulmonary arterial hypertension due to congenital heart diseases the level of homocysteine is higher and the level of endogenous sulfur dioxide is lower than in normal control children. Moreover, these biochemical parameters strongly correlated to the severity of pulmonary arterial hypertension. Authors considered homocysteine to be one of useful biochemical markers of disease severity and sulfur dioxide metabolism to be one of potential therapeutic targets in those patients.
Endogenous sulfur dioxide also has been shown to lower the proliferation rate of endothelial smooth muscle cells in blood vessels, via lowering the MAPK activity and activating adenylyl cyclase and protein kinase A. Smooth muscle cell proliferation is one of important mechanisms of hypertensive remodeling of blood vessels and their stenosis, so it is an important pathogenetic mechanism in arterial hypertension and atherosclerosis.
Endogenous sulfur dioxide in low concentrations causes endothelium-dependent vasodilation. In higher concentrations it causes endothelium-independent vasodilation and has a negative inotropic effect on cardiac output function, thus effectively lowering blood pressure and myocardial oxygen consumption. The vasodilating and bronchodilating effects of sulfur dioxide are mediated via ATP-dependent calcium channels and L-type ("dihydropyridine") calcium channels. Endogenous sulfur dioxide is also a potent antiinflammatory, antioxidant and cytoprotective agent. It lowers blood pressure and slows hypertensive remodeling of blood vessels, especially thickening of their intima. It also regulates lipid metabolism.
Endogenous sulfur dioxide also diminishes myocardial damage, caused by isoproterenol adrenergic hyperstimulation, and strengthens the myocardial antioxidant defense reserve.
As a reagent and solvent in the laboratory
Sulfur dioxide is a versatile inert solvent widely used for dissolving highly oxidizing salts. It is also used occasionally as a source of the sulfonyl group in organic synthesis. Treatment of aryl diazonium salts with sulfur dioxide and cuprous chloride yields the corresponding aryl sulfonyl chloride, for example:
As a result of its very low Lewis basicity, it is often used as a low-temperature solvent/diluent for superacids like magic acid (FSO3H/SbF5), allowing for highly reactive species like tert-butyl cation to be observed spectroscopically at low temperature (though tertiary carbocations do react with SO2 above about –30 °C, and even less reactive solvents like SO2ClF must be used at these higher temperatures).
Aspirational applications
As a refrigerant
Being easily condensed and possessing a high heat of evaporation, sulfur dioxide is a candidate material for refrigerants. Prior to the development of chlorofluorocarbons, sulfur dioxide was used as a refrigerant in home refrigerators.
Climate engineering
Injections of sulfur dioxide in the stratosphere has been proposed in climate engineering. The cooling effect would be similar to what has been observed after the large explosive 1991 eruption of Mount Pinatubo. However this form of geoengineering would have uncertain regional consequences on rainfall patterns, for example in monsoon regions.
As an air pollutant
Sulfur dioxide is a noticeable component in the atmosphere, especially following volcanic eruptions. According to the United States Environmental Protection Agency (EPA), the amount of sulfur dioxide released in the U.S. per year was:
Sulfur dioxide is a major air pollutant and has significant impacts upon human health. In addition, the concentration of sulfur dioxide in the atmosphere can influence the habitat suitability for plant communities, as well as animal life. Sulfur dioxide emissions are a precursor to acid rain and atmospheric particulates. Due largely to the US EPA's Acid Rain Program, the U.S. has had a 33% decrease in emissions between 1983 and 2002. This improvement resulted in part from flue-gas desulfurization, a technology that enables SO2 to be chemically bound in power plants burning sulfur-containing coal or oil. In particular, calcium oxide (lime) reacts with sulfur dioxide to form calcium sulfite:
CaO + SO2 → CaSO3
Aerobic oxidation of the CaSO3 gives CaSO4, anhydrite. Most gypsum sold in Europe comes from flue-gas desulfurization.
To control sulfur emissions, dozens of methods with relatively high efficiencies have been developed for fitting of coal-fired power plants.
Sulfur can be removed from coal during burning by using limestone as a bed material in fluidized bed combustion.
Sulfur can also be removed from fuels before burning, preventing formation of SO2 when the fuel is burnt. The Claus process is used in refineries to produce sulfur as a byproduct. The Stretford process has also been used to remove sulfur from fuel. Redox processes using iron oxides can also be used, for example, Lo-Cat or Sulferox.
An analysis found that 18 coal-fired power stations in the western Balkans emitted two-and-half times more sulphur dioxide than all 221 coal plants in the EU combined.
Fuel additives such as calcium additives and magnesium carboxylate may be used in marine engines to lower the emission of sulfur dioxide gases into the atmosphere.
As of 2006, China was the world's largest sulfur dioxide polluter, with 2005 emissions estimated to be . This amount represents a 27% increase since 2000, and is roughly comparable with U.S. emissions in 1980.
Safety
Inhalation
Incidental exposure to sulfur dioxide is routine, e.g. the smoke from matches, coal, and sulfur-containing fuels.
Sulfur dioxide is mildly toxic and can be hazardous in high concentrations. Long-term exposure to low concentrations is also problematic. A 2011 systematic review concluded that exposure to sulfur dioxide is associated with preterm birth.
U.S. regulations
In 2008, the American Conference of Governmental Industrial Hygienists reduced the short-term exposure limit to 0.25 parts per million (ppm). In the US, the OSHA set the PEL at 5 ppm (13 mg/m3) time-weighted average. Also in the US, NIOSH set the IDLH at 100 ppm. In 2010, the EPA "revised the primary SO2 NAAQS by establishing a new one-hour standard at a level of 75 parts per billion (ppb). EPA revoked the two existing primary standards because they would not provide additional public health protection given a one-hour standard at 75 ppb."
Ingestion
In the United States, the Center for Science in the Public Interest lists the two food preservatives, sulfur dioxide and sodium bisulfite, as being safe for human consumption except for certain asthmatic individuals who may be sensitive to them, especially in large amounts. Symptoms of sensitivity to sulfiting agents, including sulfur dioxide, manifest as potentially life-threatening trouble breathing within minutes of ingestion. Sulphites may also cause symptoms in non-asthmatic individuals, namely dermatitis, urticaria, flushing, hypotension, abdominal pain and diarrhea, and even life-threatening anaphylaxis.
See also
Bunker fuel
National Ambient Air Quality Standards
Sulfur trioxide
Sulfur–iodine cycle
References
External links
Global map of sulfur dioxide distribution
United States Environmental Protection Agency Sulfur Dioxide page
International Chemical Safety Card 0074
IARC Monographs. "Sulfur Dioxide and some Sulfites, Bisulfites and Metabisulfites". vol. 54. 1992. p. 131.
NIOSH Pocket Guide to Chemical Hazards
CDC – Sulfure Dioxide – NIOSH Workplace Safety and Health Topic
Sulfur Dioxide, Molecule of the Month
Acidic oxides
IARC Group 3 carcinogens
Industrial gases
Interchalcogens
Preservatives
Refrigerants
Airborne pollutants
Sulfur oxides
Gaseous signaling molecules
Triatomic molecules
Reducing agents
Inorganic solvents
Hypervalent molecules
E-number additives
Sulfur(IV) compounds |
null | null | Aaron Copland | eng_Latn | Aaron Copland (, ; November 14, 1900December 2, 1990) was an American composer, composition teacher, writer, and later a conductor of his own and other American music. Copland was referred to by his peers and critics as "the Dean of American Composers". The open, slowly changing harmonies in much of his music are typical of what many people consider to be the sound of American music, evoking the vast American landscape and pioneer spirit. He is best known for the works he wrote in the 1930s and 1940s in a deliberately accessible style often referred to as "populist" and which the composer labeled his "vernacular" style. Works in this vein include the ballets Appalachian Spring, Billy the Kid and Rodeo, his Fanfare for the Common Man and Third Symphony. In addition to his ballets and orchestral works, he produced music in many other genres, including chamber music, vocal works, opera and film scores.
After some initial studies with composer Rubin Goldmark, Copland traveled to Paris, where he first studied with Isidor Philipp and Paul Vidal, then with noted pedagogue Nadia Boulanger. He studied three years with Boulanger, whose eclectic approach to music inspired his own broad taste. Determined upon his return to the U.S. to make his way as a full-time composer, Copland gave lecture-recitals, wrote works on commission and did some teaching and writing. However, he found that composing orchestral music in the modernist style, which he had adopted while studying abroad, was a financially contradictory approach, particularly in light of the Great Depression. He shifted in the mid-1930s to a more accessible musical style which mirrored the German idea of ("music for use"), music that could serve utilitarian and artistic purposes. During the Depression years, he traveled extensively to Europe, Africa, and Mexico, formed an important friendship with Mexican composer Carlos Chávez and began composing his signature works.
During the late 1940s, Copland became aware that Stravinsky and other fellow composers had begun to study Arnold Schoenberg's use of twelve-tone (serial) techniques. After he had been exposed to the works of French composer Pierre Boulez, he incorporated serial techniques into his Piano Quartet (1950), Piano Fantasy (1957), Connotations for orchestra (1961) and Inscape for orchestra (1967). Unlike Schoenberg, Copland used his tone rows in much the same fashion as his tonal material—as sources for melodies and harmonies, rather than as complete statements in their own right, except for crucial events from a structural point of view. From the 1960s onward, Copland's activities turned more from composing to conducting. He became a frequent guest conductor of orchestras in the U.S. and the UK and made a series of recordings of his music, primarily for Columbia Records.
Life
Early years
Aaron Copland was born in Brooklyn, New York, on November 14, 1900. He was the youngest of five children in a Conservative Jewish family of Lithuanian origins. While emigrating from Russia to the United States, Copland's father, Harris Morris Copland (1864-1945), lived and worked in Scotland for two to three years to pay for his boat fare to the US. It was there that Copland's father may have Anglicized his surname "Kaplan" to "Copland", though Copland himself believed for many years that the change had been due to an Ellis Island immigration official when his father entered the country. Copland was however unaware until late in his life that the family name had been Kaplan, and his parents never told him this. Throughout his childhood, Copland and his family lived above his parents' Brooklyn shop, H. M. Copland's, at 628 Washington Avenue (which Aaron would later describe as "a kind of neighborhood Macy's"), on the corner of Dean Street and Washington Avenue, and most of the children helped out in the store. His father was a staunch Democrat. The family members were active in Congregation Baith Israel Anshei Emes, where Aaron celebrated his Bar Mitzvah. Not especially athletic, the sensitive young man became an avid reader and often read Horatio Alger stories on his front steps.
Copland's father had no musical interest. His mother, Sarah Mittenthal Copland (1865-1942), sang, played the piano, and arranged for music lessons for her children. Copland had four older siblings: two older brothers, Ralph Copland (1888-1952) and Leon Copland (c. 1891-?) and two older sisters, Laurine Copland (c. 1892-?) and Josephine Copland (c. 1894-?). Of his siblings, oldest brother Ralph was the most advanced musically, proficient on the violin. His sister Laurine had the strongest connection with Aaron; she gave him his first piano lessons, promoted his musical education, and supported him in his musical career. A student at the Metropolitan Opera School and a frequent opera-goer, Laurine also brought home libretti for Aaron to study. Copland attended Boys High School and in the summer went to various camps. Most of his early exposure to music was at Jewish weddings and ceremonies, and occasional family musicales.
Copland began writing songs at the age of eight and a half. His earliest notated music, about seven bars he wrote when age 11, was for an opera scenario he created and called Zenatello. From 1913 to 1917 he took piano lessons with Leopold Wolfsohn, who taught him the standard classical fare. Copland's first public music performance was at a Wanamaker's recital. By the age of 15, after attending a concert by Polish composer-pianist Ignacy Jan Paderewski, Copland decided to become a composer. At age 16, Copland heard his first symphony at the Brooklyn Academy of Music. After attempts to further his music study from a correspondence course, Copland took formal lessons in harmony, theory, and composition from Rubin Goldmark, a noted teacher and composer of American music (who had given George Gershwin three lessons). Goldmark, with whom Copland studied between 1917 and 1921, gave the young Copland a solid foundation, especially in the Germanic tradition. As Copland stated later: "This was a stroke of luck for me. I was spared the floundering that so many musicians have suffered through incompetent teaching." But Copland also commented that the maestro had "little sympathy for the advanced musical idioms of the day" and his "approved" composers ended with Richard Strauss.
Copland's graduation piece from his studies with Goldmark was a three-movement piano sonata in a Romantic style. But he had also composed more original and daring pieces which he did not share with his teacher. In addition to regularly attending the Metropolitan Opera and the New York Symphony, where he heard the standard classical repertory, Copland continued his musical development through an expanding circle of musical friends. After graduating from high school, Copland played in dance bands. Continuing his musical education, he received further piano lessons from Victor Wittgenstein, who found his student to be "quiet, shy, well-mannered, and gracious in accepting criticism". Copland's fascination with the Russian Revolution and its promise for freeing the lower classes drew a rebuke from his father and uncles. In spite of that, in his early adult life Copland would develop friendships with people with socialist and communist leanings.
Study in Paris
Copland's passion for the latest European music, plus glowing letters from his friend Aaron Schaffer, inspired him to go to Paris for further study. An article in Musical America about a summer school program for American musicians at the Fontainebleau School of Music, offered by the French government, encouraged Copland still further. His father wanted him to go to college, but his mother's vote in the family conference allowed him to give Paris a try. On arriving in France, he studied at Fontainebleau with pianist and pedagogue Isidor Philipp and composer Paul Vidal. When Copland found Vidal too much like Goldmark, he switched at the suggestion of a fellow student to Nadia Boulanger, then aged 34. He had initial reservations: "No one to my knowledge had ever before thought of studying with a woman." She interviewed him, and recalled later: "One could tell his talent immediately."
Boulanger had as many as 40 students at once and employed a formal regimen that Copland had to follow. Copland found her incisive mind much to his liking and found her ability to critique a composition impeccable. Boulanger "could always find the weak spot in a place you suspected was weak.... She could also tell you why it was weak [italics Copland]." He wrote in a letter to his brother Ralph, "This intellectual Amazon is not only professor at the Conservatoire, is not only familiar with all music from Bach to Stravinsky, but is prepared for anything worse in the way of dissonance. But make no mistake ... A more charming womanly woman never lived." Copland later wrote that "it was wonderful for me to find a teacher with such openness of mind, while at the same time she held firm ideas of right and wrong in musical matters. The confidence she had in my talents and her belief in me were at the very least flattering and more—they were crucial to my development at this time of my career." Though he planned on only one year abroad, he studied with her for three years, finding her eclectic approach inspired his own broad musical taste.
Along with his studies with Boulanger, Copland took classes in French language and history at the Sorbonne, attended plays, and frequented Shakespeare and Company, the English-language bookstore that was a gathering-place for expatriate American writers. Among this group in the heady cultural atmosphere of Paris in the 1920s were Paul Bowles, Ernest Hemingway, Sinclair Lewis, Gertrude Stein, and Ezra Pound, as well as artists like Pablo Picasso, Marc Chagall, and Amedeo Modigliani. Also influential on the new music were the French intellectuals Marcel Proust, Paul Valéry, Jean-Paul Sartre, and André Gide; the latter cited by Copland as being his personal favorite and most read. Travels to Italy, Austria, and Germany rounded out Copland's musical education. During his stay in Paris, Copland began writing musical critiques, the first on Gabriel Fauré, which helped spread his fame and stature in the music community.
1925 to 1935
After a fruitful stay in Paris, Copland returned to America optimistic and enthusiastic about the future, determined to make his way as a full-time composer. He rented a studio apartment on New York City's Upper West Side in the Empire Hotel, close to Carnegie Hall and other musical venues and publishers. He remained in that area for the next thirty years, later moving to Westchester County, New York. Copland lived frugally and survived financially with help from two $2,500 Guggenheim Fellowships in 1925 and 1926 (each of the two ). Lecture-recitals, awards, appointments, and small commissions, plus some teaching, writing, and personal loans kept him afloat in the subsequent years through World War II. Also important, especially during the Depression, were wealthy patrons who underwrote performances, helped pay for publication of works and promoted musical events and composers. Among those mentors was Serge Koussevitzky, the music director of the Boston Symphony Orchestra and known as a champion of "new music". Koussevitsky would prove to be very influential in Copland's life, and was perhaps the second most important figure in his career after Boulanger. Beginning with the Symphony for Organ and Orchestra (1924), Koussevitzky would perform more of Copland's music than that of any the composer's contemporaries, at a time when other conductors were programming only a few of Copland's works.
Soon after his return to the United States, Copland was exposed to the artistic circle of photographer Alfred Stieglitz. While Copland did not care for Stieglitz's domineering attitude, he admired his work and took to heart Stieglitz's conviction that American artists should reflect "the ideas of American Democracy". This ideal influenced not just the composer but also a generation of artists and photographers, including Paul Strand, Edward Weston, Ansel Adams, Georgia O'Keeffe, and Walker Evans. Evans' photographs inspired portions of Copland's opera The Tender Land.
In his quest to take up the slogan of the Stieglitz group, "Affirm America", Copland found only the music of Carl Ruggles and Charles Ives upon which to draw. Without what Copland called a "usable past" in American classical composers, he looked toward jazz and popular music, something he had already started to do while in Europe. In the 1920s, Gershwin, Bessie Smith, and Louis Armstrong were in the forefront of American popular music and jazz. By the end of the decade, Copland felt his music was going in a more abstract, less jazz-oriented direction. However, as large swing bands such as those of Benny Goodman and Glenn Miller became popular in the 1930s, Copland took a renewed interest in the genre.
Inspired by the example of Les Six in France, Copland sought out contemporaries such as Roger Sessions, Roy Harris, Virgil Thomson, and Walter Piston, and quickly established himself as a spokesman for composers of his generation. He also helped found the Copland-Sessions Concerts to showcase these composers' chamber works to new audiences. Copland's relationship with these men, who became known as "commando unit", was one of both support and rivalry, and he played a key role in keeping them together until after World War II. He was also generous with his time with nearly every American young composer he met during his life, later earning the title "Dean of American Music".
With the knowledge he had gained from his studies in Paris, Copland came into demand as a lecturer and writer on contemporary European classical music. From 1927 to 1930 and 1935 to 1938, he taught classes at The New School for Social Research in New York City. Eventually, his New School lectures would appear in the form of two books—What to Listen for in Music (1937, revised 1957) and Our New Music (1940, revised 1968 and retitled The New Music: 1900–1960). During this period, Copland also wrote regularly for The New York Times, The Musical Quarterly and a number of other journals. These articles would appear in 1969 as the book Copland on Music.
Copland's compositions in the early 1920s reflected the modernist attitude that prevailed among intellectuals, that the arts need be accessible to only a cadre of the enlightened and that the masses would come to appreciate their efforts over time. However, mounting troubles with the Symphonic Ode (1929) and Short Symphony (1933) caused him to rethink this approach. It was financially contradictory, particularly in the Depression. Avant-garde music had lost what cultural historian Morris Dickstein calls "its buoyant experimental edge" and the national mood toward it had changed. As biographer Howard Pollack points out,
Copland observed two trends among composers in the 1930s: first, a continuing attempt to "simplify their musical language" and, second, a desire to "make contact" with as wide an audience as possible. Since 1927, he had been in the process of simplifying, or at least paring down, his musical language, though in such a manner as to sometimes have the effect, paradoxically, of estranging audiences and performers. By 1933 ... he began to find ways to make his starkly personal language accessible to a surprisingly large number of people.
In many ways, this shift mirrored the German idea of ("music for use"), as composers sought to create music that could serve a utilitarian as well as artistic purpose. This approach encompassed two trends: first, music that students could easily learn, and second, music which would have wider appeal, such as incidental music for plays, movies, radio, etc. Toward this end, Copland provided musical advice and inspiration to The Group Theatre, a company which also attracted Stella Adler, Elia Kazan and Lee Strasberg. Philosophically an outgrowth of Stieglitz and his ideals, the Group focused on socially relevant plays by the American authors. Through it and later his work in film, Copland met several major American playwrights, including Thornton Wilder, William Inge, Arthur Miller, and Edward Albee, and considered projects with all of them.
1935 to 1950
Around 1935 Copland began to compose musical pieces for young audiences, in accordance with the first goal of American Gebrauchsmusik. These works included piano pieces (The Young Pioneers) and an opera (The Second Hurricane). During the Depression years, Copland traveled extensively to Europe, Africa, and Mexico. He formed an important friendship with Mexican composer Carlos Chávez and would return often to Mexico for working vacations conducting engagements. During his initial visit to Mexico, Copland began composing the first of his signature works, El Salón México, which he completed in 1936. In it and in The Second Hurricane Copland began "experimenting", as he phrased it, with a simpler, more accessible style. This and other incidental commissions fulfilled the second goal of American Gebrauchsmusik, creating music of wide appeal.
Concurrent with The Second Hurricane, Copland composed (for radio broadcast) "Prairie Journal" on a commission from the Columbia Broadcast System. This was one of his first pieces to convey the landscape of the American West. This emphasis on the frontier carried over to his ballet Billy the Kid (1938), which along with El Salón México became his first widespread public success. Copland's ballet music established him as an authentic composer of American music much as Stravinsky's ballet scores connected the composer with Russian music and came at an opportune time. He helped fill a vacuum for American choreographers to fill their dance repertory and tapped into an artistic groundswell, from the motion pictures of Busby Berkeley and Fred Astaire to the ballets of George Balanchine and Martha Graham, to both democratize and Americanize dance as an art form. In 1938, Copland helped form the American Composers Alliance to promote and publish American contemporary classical music. Copland was president of the organization from 1939 to 1945. In 1939, Copland completed his first two Hollywood film scores, for Of Mice and Men and Our Town, and composed the radio score "John Henry", based on the folk ballad.
While these works and others like them that would follow were accepted by the listening public at large, detractors accused Copland of pandering to the masses. Music critic Paul Rosenfeld, for one, warned in 1939 that Copland was "standing in the fork in the highroad, the two branches of which lead respectively to popular and artistic success." Even some of the composer's friends, such as composer Arthur Berger, were confused about Copland's simpler style. One, composer David Diamond, went so far as to lecture Copland: "By having sold out to the mongrel commercialists half-way already, the danger is going to be wider for you, and I beg you dear Aaron, don't sell out [entirely] yet." Copland's response was that his writing as he did and in as many genres was his response to how the Depression had affected society, as well as to new media and the audiences made available by these new media. As he himself phrased it, "The composer who is frightened of losing his artistic integrity through contact with a mass audience is no longer aware of the meaning of the word art."
The 1940s were arguably Copland's most productive years, and some of his works from this period would cement his worldwide fame. His ballet scores for Rodeo (1942) and Appalachian Spring (1944) were huge successes. His pieces Lincoln Portrait and Fanfare for the Common Man became patriotic standards. Also important was the Third Symphony. Composed in a two-year period from 1944 to 1946, it became Copland's best-known symphony. The Clarinet Concerto (1948), scored for solo clarinet, strings, harp, and piano, was a commission piece for bandleader and clarinetist Benny Goodman and a complement to Copland's earlier jazz-influenced work, the Piano Concerto (1926). His Four Piano Blues is an introspective composition with a jazz influence. Copland finished the 1940s with two film scores, one for William Wyler's The Heiress and one for the film adaptation of John Steinbeck's novel The Red Pony.
In 1949, Copland returned to Europe, where he found French composer Pierre Boulez dominating the group of post-war avant-garde composers there. He also met with proponents of twelve-tone technique, based on the works of Arnold Schoenberg, and found himself interested in adapting serial methods to his own musical voice.
1950s and 1960s
In 1950, Copland received a U.S.-Italy Fulbright Commission scholarship to study in Rome, which he did the following year. Around this time, he also composed his Piano Quartet, adopting Schoenberg's twelve-tone method of composition, and Old American Songs (1950), the first set of which was premiered by Peter Pears and Benjamin Britten, the second by William Warfield. During the 1951–52 academic year, Copland gave a series of lectures under the Charles Eliot Norton Professorship at Harvard University. These lectures were published as the book Music and Imagination.
Because of his leftist views, which had included his support of the Communist Party USA ticket during the 1936 presidential election and his strong support of Progressive Party candidate Henry A. Wallace during the 1948 presidential election, Copland was investigated by the FBI during the Red scare of the 1950s. He was included on an FBI list of 151 artists thought to have Communist associations and found himself blacklisted, with A Lincoln Portrait withdrawn from the 1953 inaugural concert for President Eisenhower. Called later that year to a private hearing at the United States Capitol in Washington, D.C., Copland was questioned by Joseph McCarthy and Roy Cohn about his lecturing abroad and his affiliations with various organizations and events. In the process, McCarthy and Cohn neglected completely Copland's works, which made a virtue of American values. Outraged by the accusations, many members of the musical community held up Copland's music as a banner of his patriotism. The investigations ceased in 1955 and were closed in 1975.
The McCarthy probes did not seriously affect Copland's career and international artistic reputation, taxing of his time, energy, and emotional state as they might have been. Nevertheless, beginning in 1950, Copland—who had been appalled at Stalin's persecution of Shostakovich and other artists—began resigning from participation in leftist groups. Copland, Pollack states, "stayed particularly concerned about the role of the artist in society". He decried the lack of artistic freedom in the Soviet Union, and in his 1954 Norton lecture he asserted that loss of freedom under Soviet Communism deprived artists of "the immemorial right of the artist to be wrong". He began to vote Democratic, first for Stevenson and then for Kennedy.
Potentially more damaging for Copland was a sea-change in artistic tastes, away from the Populist mores that infused his work of the 1930s and 40s. Beginning in the 1940s, intellectuals assailed Popular Front culture, to which Copland's music was linked, and labeled it, in Dickstein's words, as "hopelessly middlebrow, a dumbing down of art into toothless entertainment". They often linked their disdain for Populist art with technology, new media and mass audiences—in other words, the areas of radio, television and motion pictures, for which Copland either had or soon would write music, as well as his popular ballets. While these attacks actually began at the end of the 1930s with the writings of Clement Greenberg and Dwight Macdonald for Partisan Review, they were based in anti-Stalinist politics and would accelerate in the decades following World War II.
Despite any difficulties that his suspected Communist sympathies might have posed, Copland traveled extensively during the 1950s and early 60s to observe the avant-garde styles of Europe, hear compositions by Soviet composers not well known in the West and experience the new school of Polish music. While in Japan, he was taken with the work of Tōru Takemitsu and began a correspondence with him that would last over the next decade. Copland revised his text "The New Music" with comments on the styles that he encountered. He found much of what he heard dull and impersonal. Electronic music seemed to have "a depressing sameness of sound," while aleatoric music was for those "who enjoy teetering on the edge of chaos". As he summarized, "I've spent most of my life trying to get the right note in the right place. Just throwing it open to chance seems to go against my natural instincts."
In 1952, Copland received a commission from the League of Composers, funded by a grant from Richard Rodgers and Oscar Hammerstein, to write an opera for television. While Copland was aware of the potential pitfalls of that genre, which included weak libretti and demanding production values, he had also been thinking about writing an opera since the 1940s. Among the subjects he had considered were Theodore Dreiser's An American Tragedy and Frank Norris's McTeague He finally settled on James Agee's Let Us Now Praise Famous Men, which seemed appropriate for the more intimate setting of television and could also be used in the "college trade", with more schools mounting operas than they had before World War II. The resulting opera, The Tender Land, was written in two acts but later expanded to three. As Copland feared, when the opera premiered in 1954 critics found the libretto to be weak. In spite of its flaws, the opera became one of the few American operas to enter the standard repertory.
In 1957, 1958, and 1976, Copland was the Music Director of the Ojai Music Festival, a classical and contemporary music festival in Ojai, California. For the occasion of the Metropolitan Museum of Art Centennial, Copland composed Ceremonial Fanfare For Brass Ensemble to accompany the exhibition "Masterpieces Of Fifty Centuries". Leonard Bernstein, Piston, William Schuman, and Thomson also composed pieces for the Museum's Centennial exhibitions.
Later years
From the 1960s, Copland turned increasingly to conducting. Though not enamored with the prospect, he found himself without new ideas for composition, saying, "It was exactly as if someone had simply turned off a faucet." He became a frequent guest conductor in the United States and the United Kingdom and made a series of recordings of his music, primarily for Columbia Records. In 1960, RCA Victor released Copland's recordings with the Boston Symphony Orchestra of the orchestral suites from Appalachian Spring and The Tender Land; these recordings were later reissued on CD, as were most of Copland's Columbia recordings (by Sony).
From 1960 to his death, Copland resided at Cortlandt Manor, New York. Known as Rock Hill, his home was added to the National Register of Historic Places in 2003 and further designated a National Historic Landmark in 2008. Copland's health deteriorated through the 1980s, and he died of Alzheimer's disease and respiratory failure on December 2, 1990, in North Tarrytown, New York (now Sleepy Hollow). Following his death, his ashes were scattered over the Tanglewood Music Center near Lenox, Massachusetts. Much of his large estate was bequeathed to the creation of the Aaron Copland Fund for Composers, which bestows over $600,000 per year to performing groups.
Personal life
Copland never enrolled as a member of any political party. Nevertheless, he inherited a considerable interest in civic and world events from his father. His views were generally progressive and he had strong ties with numerous colleagues and friends in the Popular Front, including Odets. Early in his life, Copland developed, in Pollack's words, "a deep admiration for the works of Frank Norris, Theodore Dreiser and Upton Sinclair, all socialists whose novels passionately excoriated capitalism's physical and emotional toll on the average man". Even after the McCarthy hearings, he remained a committed opponent of militarism and the Cold War, which he regarded as having been instigated by the United States. He condemned it as "almost worse for art than the real thing". Throw the artist "into a mood of suspicion, ill-will, and dread that typifies the cold war attitude and he'll create nothing".
While Copland had various encounters with organized religious thought, which influenced some of his early compositions, he remained agnostic. He was close with Zionism during the Popular Front movement, when it was endorsed by the left. Pollack writes,
Like many contemporaries, Copland regarded Judaism alternately in terms of religion, culture, and race; but he showed relatively little involvement in any aspect of his Jewish heritage.... At the same time, he had ties to Christianity, identifying with such profoundly Christian writers as Gerard Manley Hopkins and often spending Christmas Day at home with a special dinner with close friends.... In general, his music seemed to evoke Protestant hymns as often as it did Jewish chant....Copland characteristically found connections among various religious traditions.... But if Copland was discreet about his Jewish background, he never hid it, either.
Pollack states that Copland was gay and that the composer came to an early acceptance and understanding of his sexuality. Like many at that time, Copland guarded his privacy, especially in regard to his homosexuality. He provided few written details about his private life and even after the Stonewall riots of 1969, showed no inclination to "come out". However, he was one of the few composers of his stature to live openly and travel with his intimates. They tended to be talented, younger men involved in the arts, and the age-gap between them and the composer widened as he grew older. Most became enduring friends after a few years and, in Pollack's words, "remained a primary source of companionship". Among Copland's love affairs were ones with photographer Victor Kraft, artist Alvin Ross, pianist Paul Moor, dancer Erik Johns, composer John Brodbin Kennedy, and painter Prentiss Taylor.
Victor Kraft became a constant in Copland's life, though their romance might have ended by 1944. Originally a violin prodigy when the composer met him in 1932, Kraft gave up music to pursue a career in photography, in part due to Copland's urging. Kraft would leave and re-enter Copland's life, often bringing much stress with him as his behavior became increasingly erratic, sometimes confrontational. Kraft fathered a child to whom Copland later provided financial security, through a bequest from his estate.
Music
Vivian Perlis, who collaborated with Copland on his autobiography, writes: "Copland's method of composing was to write down fragments of musical ideas as they came to him. When he needed a piece, he would turn to these ideas (his 'gold nuggets')." if one or more of these nuggets looked promising, he would then write a piano sketch and eventually work on them at the keyboard. The piano, Perlis writes, "was so integral to his composing that it permeated his compositional style, not only in the frequent use in the instrument but in more subtle and complex ways". His habit of turning to the keyboard tended to embarrass Copland until he learned that Stravinsky also did so.
Copland would not consider the specific instrumentation for a piece until it was complete and notated. Nor, according to Pollack, did he generally work in linear fashion, from beginning to end of a composition. Instead, he tended to compose whole sections in no particular order and surmise their eventual sequence after all those parts were complete, much like assembling a collage. Copland himself admitted, "I don't compose. I assemble materials." Many times, he included material he had written years earlier. If the situation dictated, as it did with his film scores, Copland could work quickly. Otherwise, he tended to write slowly whenever possible. Even with this deliberation, Copland considered composition, in his words, "the product of the emotions", which included "self-expression" and "self-discovery".
Influences
While Copland's earliest musical inclinations as a teenager ran toward Chopin, Debussy, Verdi and the Russian composers, Copland's teacher and mentor Nadia Boulanger became his most important influence. Copland especially admired Boulanger's total grasp of all classical music, and he was encouraged to experiment and develop a "clarity of conception and elegance in proportion". Following her model, he studied all periods of classical music and all forms—from madrigals to symphonies. This breadth of vision led Copland to compose music for numerous settings—orchestra, opera, solo piano, small ensemble, art song, ballet, theater and film. Boulanger particularly emphasized "la grande ligne" (the long line), "a sense of forward motion ... the feeling for inevitability, for the creating of an entire piece that could be thought of as a functioning entity".
During his studies with Boulanger in Paris, Copland was excited to be so close to the new post-Impressionistic French music of Ravel, Roussel, and Satie, as well as Les Six, a group that included Milhaud, Poulenc, and Honegger. Webern, Berg, and Bartók also impressed him. Copland was "insatiable" in seeking out the newest European music, whether in concerts, score reading or heated debate. These "moderns" were discarding the old laws of composition and experimenting with new forms, harmonies and rhythms, and including the use of jazz and quarter-tone music. Milhaud was Copland's inspiration for some of his earlier "jazzy" works. He was also exposed to Schoenberg and admired his earlier atonal pieces, thinking Schoenberg's Pierrot lunaire above all others. Copland named Igor Stravinsky as his "hero" and his favorite 20th-century composer. Copland especially admired Stravinsky's "jagged and uncouth rhythmic effects", "bold use of dissonance", and "hard, dry, crackling sonority".
Another inspiration for much of Copland's music was jazz. Although familiar with jazz back in America—having listened to it and also played it in bands—he fully realized its potential while traveling in Austria: "The impression of jazz one receives in a foreign country is totally unlike the impression of such music heard in one's own country ... when I heard jazz played in Vienna, it was like hearing it for the first time." He also found that the distance from his native country helped him see the United States more clearly. Beginning in 1923, he employed "jazzy elements" in his classical music, but by the late 1930s, he moved on to Latin and American folk tunes in his more successful pieces. Although his early focus of jazz gave way to other influences, Copland continued to make use of jazz in more subtle ways in later works.
Copland's work from the late 1940s onward included experimentation with Schoenberg's twelve-tone system, resulting in two major works, the Piano Quartet (1950) and the Piano Fantasy (1957).
Early works
Copland's compositions before leaving for Paris were mainly short works for piano and art songs, inspired by Liszt and Debussy. In them, he experimented with ambiguous beginnings and endings, rapid key changes, and the frequent use of tritones. His first published work, The Cat and the Mouse (1920), was a piece for piano solo based on the Jean de la Fontaine fable. In Three Moods (1921), Copland's final movement is entitled "Jazzy", which he noted "is based on two jazz melodies and ought to make the old professors sit up and take notice".
The Symphony for Organ and Orchestra established Copland as a serious modern composer. Musicologist Gayle Murchison cites Copland's use melodic, harmonic and rhythmic elements endemic in jazz, which he would also use in his Music for the Theater and Piano Concerto to evoke an essentially "American" sound. he fuses these qualities with modernist elements such as octatonic and whole-tone scales, polyrhythmic ostinato figures, and dissonant counterpoint. Murchinson points out the influence of Igor Stravinsky in the work's nervous, driving rhythms and some of its harmonic language. Copland in hindsight found the work too "European" as he consciously sought a more consciously American idiom to evoke in his future work.
Visits to Europe in 1926 and 1927 brought him into contact with the most recent developments there, including Webern's Five Pieces for Orchestra, which greatly impressed him. In August 1927, while staying in Königstein, Copland wrote Poet's Song, a setting of a text by E. E. Cummings and his first composition using Schoenberg's twelve-tone technique. This was followed by the Symphonic Ode (1929) and the Piano Variations (1930), both of which rely on the exhaustive development of a single short motif. This procedure, which provided Copland with more formal flexibility and a greater emotional range than in his earlier music, is similar to Schoenberg's idea of "continuous variation" and, according to Copland's own admission, was influenced by the twelve-tone method, though neither work actually uses a twelve-tone row.
The other major work of Copland's first period is the Short Symphony (1933). In it, music critic and musicologist Michael Steinberg writes, the "jazz-influenced dislocations of meter that are so characteristic of Copland's music of the 1920s are more prevalent than ever". Compared to the Symphonic Ode, the orchestration is much leaner and the composition itself more concentrated. In its combination and refinement of modernist and jazz elements, Steinberg calls the Short Symphony "a remarkable synthesis of the learned and the vernacular, and thus, in all its brevity [the work last just 15 minutes], a singularly 'complete' representation of its composer". However, Copland moved from this work toward more accessible works and folk sources.
Populist works
Copland wrote El Salón México between 1932 and 1936, which met with a popular acclaim that contrasted the relative obscurity of most of his previous works. Inspiration for this work came from Copland's vivid recollection of visiting the "Salon Mexico" dancehall where he witnessed a more intimate view of Mexico's nightlife. Copland derived his melodic material for this piece freely from two collections of Mexican folk tunes, changing pitches and varying rhythms. The use of a folk tune with variations set in a symphonic context started a pattern he repeated in many of his most successful works right on through the 1940s. It also marked a shift in emphasis from a unified musical structure to the rhetorical effect the music might have on an audience and showed Copland refining a simplified, more accessible musical language.
El Salón prepared Copland to write the ballet score Billy the Kid, which became, in Pollack's words, an "archetypical depiction of the legendary American West". Based on a Walter Noble Burns novel, with choreography by Eugene Loring, Billy was among the first to display an American music and dance vocabulary. Copland used six cowboy folk songs to provide period atmosphere and employed polyrhythm and polyharmony when not quoting these tunes literally to maintain the work's overall tone. In this way, Copland's music worked much in the same way as the murals of Thomas Hart Benton, in that it employed elements that could be grasped easily by a mass audience. The ballet premiered in New York in 1939, with Copland recalling: "I cannot remember another work of mine that was so unanimously received." Along with the ballet Rodeo, Billy the Kid became, in the words of musicologist Elizabeth Crist, "the basis for Copland's reputation as a composer of Americana" and defines "an uncomplicated form of American nationalism".
Copland's brand of nationalism in his ballets differed from that of European composers such as Béla Bartók, who tried to preserve the folk tones they used as close to the original as possible. Copland enhanced the tunes he used with contemporary rhythms, textures and structures. In what could seem contradictory, he used complex harmonies and rhythms to simplify folk melodies and make them more accessible and familiar to his listeners. Except for the Shaker tune in Appalachian Spring, Copland often syncopates traditional melodies, changes their metric patterns and note values. In Billy the Kid, he derives many of the work's sparse harmonies from the implied harmonic constructions of the cowboy tunes themselves.
Like Stravinsky, Copland mastered the ability to create a coherent, integrated composition from what was essentially a mosaic of divergent folk-based and original elements. In that sense, Copland's Populist works such as Billy the Kid, Rodeo, Appalachian Spring are not far removed from Stravinsky's ballet The Rite of Spring. Within that framework, however, Copland preserved the American atmosphere of these ballets through what musicologist Elliott Antokoletz calls "the conservative handling of open diatonic sonorities", which fosters "a pastoral quality" in the music. This is especially true in the opening of Appalachian Spring, where the harmonizations remain "transparent and bare, suggested by the melodic disposition of the Shaker tune". Variations which contrast to this tune in rhythm, key, texture and dynamics, fit within Copland's compositional practice of juxtaposing structural blocks.
Film scores
When Hollywood beckoned concert hall composers in the 1930s with promises of better films and higher pay, Copland saw both a challenge for his abilities as a composer as well as an opportunity to expand his reputation and audience for his more serious works. In a departure from other film scores of the time, Copland's work largely reflected his own style, instead of the usual borrowing from the late-Romantic period. He often avoided the full orchestra, and he rejected the common practice of using a leitmotif to identify characters with their own personal themes. He instead matched a theme to the action, while avoiding the underlining of every action with exaggerated emphasis. Another technique Copland employed was to keep silent during intimate screen moments and only begin the music as a confirming motive toward the end of a scene. Virgil Thomson wrote that the score for Of Mice and Men established "the most distinguished populist musical style yet created in America". Many composers who scored for western movies, particularly between 1940 and 1960, were influenced by Copland's style, though some also followed the late Romantic "Max Steiner" approach, which was considered more conventional and desirable.
Later works
Copland's work in the late 1940s and 1950s included use of Schoenberg's twelve-tone system, a development that he had recognized but not fully embraced. He had also believed the atonality of serialized music to run counter to his desire to reach a wide audience. Copland therefore approached dodecaphony with some initial skepticism. While in Europe in 1949, he heard a number of serial works but did not admire much of it because "so often it seemed that individuality was sacrificed to the method". The music of French composer Pierre Boulez showed Copland that the technique could be separated from the "old Wagnerian" aesthetic with which he had associated it previously. Subsequent exposure to the late music of Austrian composer Anton Webern and twelve-tone pieces by Swiss composer Frank Martin and Italian composer Luigi Dallapiccola strengthened this opinion.
Copland came to the conclusion that composing along serial lines was "nothing more than an angle of vision. Like fugal treatment, it is a stimulus that enlivens musical thinking, especially when applied to a series of tones that lend themselves to that treatment." He began his first serial work, the "Piano Fantasy", in 1951 to fulfill a commission from the young virtuoso pianist William Kapell. The piece became one of his most challenging works, over which he labored until 1957. During the work's development, in 1953, Kapell died in an aircraft crash. Critics lauded the "Fantasy" when it was finally premiered, calling the piece "an outstanding addition to his own oeuvre and to contemporary piano literature" and "a tremendous achievement". Jay Rosenfield stated: "This is a new Copland to us, an artist advancing with strength and not building on the past alone."
Serialism allowed Copland a synthesis of serial and non-serial practices. Before he did this, according to musicologist Joseph Straus, the philosophical and compositional difference between non-tonal composers such as Schoenberg and tonal composers like Stravinsky had been considered too wide a gulf to bridge. Copland wrote that, to him, serialism pointed in two opposite directions, one "toward the extreme of total organization with electronic applications" and the other "a gradual absorption into what had become a very freely interpreted tonalism [italics Copland]". The path he said he chose was the latter one, which he said, when he described his Piano Fantasy, allowed him to incorporate "elements able to be associated with the twelve-tone method and also with music tonally conceived". This practice differed markedly from Schoenberg, who used his tone rows as complete statements around which to structure his compositions. Copland used his rows not very differently from how he fashioned the material in his tonal pieces. He saw his rows as sources for melodies and harmonies, not as complete and independent entities, except at points in the musical structure that dictated the complete statement of a row.
Even after Copland started using 12-tone techniques, he did not stick to them exclusively but went back and forth between tonal and non-tonal compositions. Other late works include: Dance Panels (1959, ballet music), Something Wild (1961, his last film score, much of which would be later incorporated into his Music for a Great City), Connotations (1962, for the new Lincoln Center Philharmonic hall), Emblems (1964, for wind band), Night Thoughts (1972, for the Van Cliburn International Piano Competition), and Proclamation (1982, his last work, started in 1973).
Critic, writer, teacher
Copland did not consider himself a professional writer. He called his writing "a byproduct of my trade" as "a kind of salesman for contemporary music". As such, he wrote prolifically about music, including pieces on music criticism analysis, on musical trends, and on his own compositions. An avid lecturer and lecturer-performer, Copland eventually collected his presentation notes into three books, What to Listen for in Music (1939), Our New Music (1941), and Music and Imagination (1952). In the 1980s, he collaborated with Vivian Perlis on a two-volume autobiography, Copland: 1900 Through 1942 (1984) and Copland Since 1943 (1989). Along with the composer's first-person narrative, these two books incorporate 11 "interludes" by Perlis and other sections from friends and peers. Some controversy arose over the second volume's increased reliance over the first on old documents for source material. Due to the then-advanced stage of Copland's Alzheimer's and the resulting memory loss, however, this fallback to previous material was inevitable. The use in both books of letters and other unpublished sources, expertly researched and organized, made them what Pollack terms "invaluable".
During his career, Copland met and helped hundreds of young composers, whom he met and who were drawn to him by his continual interest and acuity into the contemporary musical scene. This assistance came mainly outside an institutional framework—other than his summers at the Berkshire Music Center at Tanglewood, a decade of teaching and curating at The New School, and a few semesters at Harvard and the State University of New York at Buffalo, Copland operated outside an academic setting. Pollack writes: "Those composers who actually studied with him were small in number and did so for only brief periods; rather, Copland helped younger composers more informally, with intermittent advice and aid." This advice included focusing on expressive content rather than on purely technical points and on developing a personal style.
Copland's willingness to foster talent extended to critiquing scores in progress that were presented to him by his peers. Composer William Schuman writes: "As a teacher, Aaron was extraordinary.... Copland would look at your music and try to understand what you were after [italics Schuman]. He didn't want to turn you into another Aaron Copland.... When he questioned something, it was in a manner that might make you want to question it yourself. Everything he said was helpful in making a younger composer realize the potential of a particular work. On the other hand, Aaron could be strongly critical."
Conductor
Although Copland studied conducting in Paris in 1921, he remained essentially a self-taught conductor with a very personal style. Encouraged by Igor Stravinsky to master conducting and perhaps emboldened by Carlos Chavez's efforts in Mexico, he began to direct his own works on his international travels in the 1940s. By the 1950s, he was also conducting the works of other composers, and after a televised appearance where he directed the New York Philharmonic, Copland became in high demand. He placed a strong emphasis in his programs on 20th-century music and lesser-known composers, and until the 1970s rarely planned concerts to feature his music exclusively. Performers and audiences generally greeted his conducting appearances as positive opportunities to hear his music as the composer intended. His efforts on behalf of other composers could be penetrating but also uneven.
Understated on the podium, Copland modeled his style after other composer/conductors such as Stravinsky and Paul Hindemith. Critics wrote of his precision and clarity before an orchestra. Observers noted that he had "none of the typical conductorial vanities". Copland's unpretentious charm was appreciated by professional musicians but some criticized his "unsteady" beat and "unexciting" interpretations. Koussevitzky advised him to "stay home and compose". Copland at times asked for conducting advice from Bernstein, who occasionally joked that Copland could conduct his works "a little better." Bernstein also noted that Copland improved over time, and he considered him a more natural conductor than Stravinsky or Hindemith. Eventually, Copland recorded nearly all his orchestral works with himself conducting.
Legacy
Copland wrote a total of about 100 works which covered a diverse range of genres. Many of these compositions, especially orchestral pieces, have remained part of the standard American repertoire. According to Pollack, Copland "had perhaps the most distinctive and identifiable musical voice produced by this country so far, an individuality ... that helped define for many what American concert music sounds like at its most characteristic and that exerted enormous influence on multitudes of contemporaries and successors." His synthesis of influences and inclinations helped create the "Americanism" of his music. The composer himself pointed out, in summarizing the American character of his music, "the optimistic tone", "his love of rather large canvases", "a certain directness in expression of sentiment", and "a certain songfulness".
While "Copland's musical rhetoric has become iconic" and "has functioned as a mirror of America," conductor Leon Botstein suggests that the composer "helped define the modern consciousness of America's ideals, character and sense of place. The notion that his music played not a subsidiary but a central role in the shaping of the national consciousness makes Copland uniquely interesting, for the historian as well as the musician." Composer Ned Rorem states, "Aaron stressed simplicity: Remove, remove, remove what isn't needed.... Aaron brought leanness to America, which set the tone for our musical language throughout [World War II]. Thanks to Aaron, American music came into its own."
Awards
On September 14, 1964, Aaron Copland was presented with the Presidential Medal of Freedom by President Lyndon B. Johnson.
In honor of Copland's vast influence on American music, on December 15, 1970, he was awarded the prestigious University of Pennsylvania Glee Club Award of Merit. Beginning in 1964, this award "established to bring a declaration of appreciation to an individual each year that has made a significant contribution to the world of music and helped to create a climate in which our talents may find valid expression".
Copland was awarded the New York Music Critics' Circle Award and the Pulitzer Prize in composition for Appalachian Spring. His scores for Of Mice and Men (1939), Our Town (1940), and The North Star (1943) all received Academy Award nominations, while The Heiress won Best Music in 1950.
In 1961, Aaron Copland was awarded the Edward MacDowell Medal by the MacDowell Colony where he was a fellow eight times (1925, 1928, 1935, 1938, 1946, 1950, 1952, 1956.)
He was a recipient of Yale University's Sanford Medal.
In 1986, he was awarded the National Medal of Arts.
He was awarded a special Congressional Gold Medal by the United States Congress in 1987.
He was made an honorary member of the Alpha Upsilon chapter of Phi Mu Alpha Sinfonia in 1961 and was awarded the fraternity's Charles E. Lutton Man of Music Award in 1970.
Notable students
In popular culture
Aaron Copland's music has served as the inspiration for a number of popular modern works of music:
"Hoedown" – Annie Moses Band
"Fanfare for the Common Man" – Emerson, Lake & Palmer
"The Greatest Man That Ever Lived (Variations on a Shaker Hymn)" – Weezer (partially based upon "Variations on a Shaker Hymn")
Copland's music was prominently featured throughout Spike Lee's 1998 film, He Got Game.
Selected works
Scherzo Humoristique: The Cat and the Mouse (1920)
Four Motets (1921)
Three Moods (piano solo) (1921)
Passacaglia (piano solo) (1922)
Symphony for Organ and Orchestra (1924)
Music for the Theater (1925)
Concerto for Piano and Orchestra (1926)
Symphonic Ode (1927–1929)
Piano Variations (1930)
Grohg (ballet) (1925/32)
Dance Symphony (1929) (using music from Grohg)
Short Symphony (Symphony No. 2) (1931–33)
Statements for Orchestra (1932–35)
The Second Hurricane, play-opera for high school performance (1936)
El Salón México (1936)
Billy the Kid (ballet) (1938)
Quiet City (1940)
Our Town (1940)
Piano Sonata (1939–41)
An Outdoor Overture, written for high school orchestras (1938) and transcribed for wind band (1941)
Fanfare for the Common Man (1942)
Lincoln Portrait (1942)
Rodeo (ballet) (1942)
Danzón cubano (1942)
Music for the Movies (1942)
Sonata for violin and piano (1943)
Appalachian Spring (ballet) (1944)
Third Symphony (1944–1946)
In the Beginning (1947)
The Red Pony (1948)
Clarinet Concerto (commissioned by Benny Goodman) (1947–1948)
Twelve Poems of Emily Dickinson (1950)
Piano Quartet (1950)
Old American Songs (Book One 1950, Book Two 1952)
The Tender Land (opera) (1954)
Canticle of Freedom (1955)
Orchestral Variations (orchestration of Piano Variations) (1957)
Piano Fantasy (1957)
Dance Panels (ballet) (1959; revised 1962)
Connotations (1962)
Down A Country Lane (1962)
Music for a Great City (1964) (based on his score of the 1961 film Something Wild)
Emblems, for wind band (1964); orchestral transcription by D. Wilson Ochoa (2006)
Inscape (1967)
Duo for flute and piano (1971)
Night-Thoughts (1972)
Three Latin American Sketches (1972)
Source:
Film
Aaron Copland: A Self-Portrait (1985). Directed by Allan Miller. Biographies in Music series. Princeton, New Jersey: The Humanities.
Appalachian Spring (1996). Directed by Graham Strong, Scottish Television Enterprises. Princeton, New Jersey: Films for the Humanities.
Copland Portrait (1975). Directed by Terry Sanders, United States Information Agency. Santa Monica, California: American Film Foundation.
Fanfare for America: The Composer Aaron Copland (2001). Directed by Andreas Skipis. Produced by Hessischer Rundfunk in association with Reiner Moritz Associates. Princeton, New Jersey: Films for the Humanities & Sciences.
Written works
Copland, Aaron (1939; revised 1957), What to Listen for in Music, New York: McGraw-Hill, reprinted many times.
—— (1941; revised 1968), Our New Music (The New Music: 1900–1960, rev.), New York: W. W. Norton.
—— (1953), Music and Imagination, Harvard University Press.
—— (1960), Copland on Music, New York: Doubleday.
—— (2006). Music and Imagination, Cambridge, Massachusetts: Harvard University Press. .
References
Citations
Bibliography
Further reading
Grout, Donald Jay, and Claude V. Palisca (1996). A History of Western Music, 5th edition. New York & London: W. W. Norton. (cloth); (pbk).
External links
The Aaron Copland Collection and the Aaron Copland Collection at the Library of Congress
A Tribute to Aaron Copland at American Music Preservation.com
Aaron Copland Oral History collection at Oral History of American Music
Listening
Hoedown – Annie Moses Band
Art of the States: Aaron Copland
Audio (.ram files) of a 1961 interview for the BBC (archive from March 15, 2012; accessed June 30, 2016)
Audio (.smil files) of a 1980 interview for NPR
Fanfare for America (video)
1900 births
1990 deaths
20th-century American composers
20th-century American conductors (music)
20th-century American male musicians
20th-century American male writers
20th-century classical composers
20th-century classical pianists
20th-century jazz composers
20th-century LGBT people
American agnostics
American classical composers
American classical pianists
American expatriates in France
American film score composers
American gay musicians
American gay writers
American jazz composers
American jazz pianists
American male classical composers
American male classical pianists
American male conductors (music)
American male film score composers
American male jazz composers
American male jazz musicians
American male pianists
American music educators
American musical theatre composers
American opera composers
American people of Lithuanian-Jewish descent
American people of Russian-Jewish descent
ASCAP composers and authors
Ballet composers
Best Original Music Score Academy Award winners
Boys High School (Brooklyn) alumni
Broadway composers and lyricists
Classical musicians from New York (state)
Columbia Records artists
Composers for piano
Composers for trombone
Composers from New York City
Congressional Gold Medal recipients
Danzón composers
Deaths from Alzheimer's disease
Deaths from respiratory failure
Educators from New York City
Grammy Award winners
Hollywood blacklist
Honorary Members of the Royal Philharmonic Society
Jazz-influenced classical composers
Jazz musicians from New York (state)
Jewish agnostics
Jewish American classical composers
Jewish American classical musicians
Jewish American film score composers
Jewish American jazz composers
Jewish American writers
Jewish classical pianists
Jewish jazz musicians
Kennedy Center honorees
LGBT classical composers
LGBT classical musicians
LGBT film score composers
LGBT jazz composers
LGBT Jews
LGBT people from New York (state)
LGBT writers from the United States
MacDowell Colony fellows
Male musical theatre composers
Male opera composers
Modernist composers
Music & Arts artists
Musicians from Brooklyn
Neurological disease deaths in New York (state)
People from Cortlandt Manor, New York
People from the Upper West Side
Presidential Medal of Freedom recipients
Presidents of the American Academy of Arts and Letters
Pulitzer Prize for Music winners
Pupils of Isidor Philipp
Pupils of Paul Vidal
Pupils of Rubin Goldmark
Twelve-tone and serial composers
United States National Medal of Arts recipients
Writers from Brooklyn |
null | null | Observational learning | eng_Latn | Observational learning is learning that occurs through observing the behavior of others. It is a form of social learning which takes various forms, based on various processes. In humans, this form of learning seems to not need reinforcement to occur, but instead, requires a social model such as a parent, sibling, friend, or teacher with surroundings. Particularly in childhood, a model is someone of authority or higher status in an environment. In animals, observational learning is often based on classical conditioning, in which an instinctive behavior is elicited by observing the behavior of another (e.g. mobbing in birds), but other processes may be involved as well.
Human observational learning
Many behaviors that a learner observes, remembers, and imitates are actions that models display and display modeling, even though the model may not intentionally try to instill a particular behavior. A child may learn to swear, smack, smoke, and deem other inappropriate behavior acceptable through poor modeling. Albert Bandura claims that children continually learn desirable and undesirable behavior through observational learning. Observational learning suggests that an individual's environment, cognition, and behavior all incorporate and ultimately determine how the individual functions and models.
Through observational learning, individual behaviors can spread across a culture through a process called diffusion chain. This basically occurs when an individual first learns a behavior by observing another individual and that individual serves as a model through whom other individuals learn the behavior, and so on.
Culture plays a role in whether observational learning is the dominant learning style in a person or community. Some cultures expect children to actively participate in their communities and are therefore exposed to different trades and roles on a daily basis. This exposure allows children to observe and learn the different skills and practices that are valued in their communities.
Albert Bandura, who is known for the classic Bobo doll experiment, identified this basic form of learning in 1961. The importance of observational learning lies in helping individuals, especially children, acquire new responses by observing others' behavior.
Albert Bandura states that people's behavior could be determined by their environment. Observational learning occurs through observing negative and positive behaviors. Bandura believes in reciprocal determinism in which the environment can influence people's behavior and vice versa. For instance, the Bobo doll experiment shows that the model, in a determined environment, affects children's behavior. In this experiment Bandura demonstrates that one group of children placed in an aggressive environment would act the same way, while the control group and the other group of children placed in a passive role model environment hardly shows any type of aggression.
In communities where children's primary mode of learning is through observation, the children are rarely separated from adult activities. This incorporation into the adult world at an early age allows children to use observational learning skills in multiple spheres of life. This learning through observation requires keen attentive abilities. Culturally, they learn that their participation and contributions are valued in their communities. This teaches children that it is their duty, as members of the community, to observe others' contributions so they gradually become involved and participate further in the community.
Influential stages and factors
The stages of observational learning include exposure to the model, acquiring the model's behaviour and accepting it as one's own.
Bandura's social cognitive learning theory states that there are four factors that influence observational learning:
Attention: Observers cannot learn unless they pay attention to what's happening around them. This process is influenced by characteristics of the model, such as how much one likes or identifies with the model, and by characteristics of the observer, such as the observer's expectations or level of emotional arousal.
Retention/Memory: Observers must not only recognize the observed behavior but also remember it at some later time. This process depends on the observer's ability to code or structure the information in an easily remembered form or to mentally or physically rehearse the model's actions.
Initiation/Motor: Observers must be physically and/intellectually capable of producing the act. In many cases, the observer possesses the necessary responses. But sometimes, reproducing the model's actions may involve skills the observer has not yet acquired. It is one thing to carefully watch a circus juggler, but it is quite another to go home and repeat those acts.
Motivation: The observer must have motivation to recreate the observed behavior.
Bandura clearly distinguishes between learning and performance. Unless motivated, a person does not produce learned behavior. This motivation can come from external reinforcement, such as the experimenter's promise of reward in some of Bandura's studies, or the bribe of a parent. Or it can come from vicarious reinforcement, based on the observation that models are rewarded. High-status models can affect performance through motivation. For example, girls aged 11 to 14 performed better on a motor performance task when they thought it was demonstrated by a high-status cheerleader than by a low-status model.
Some have even added a step between attention and retention involving encoding a behavior.
Observational learning leads to a change in an individual's behavior along three dimensions:
An individual thinks about a situation in a different way and may have incentive to react to it.
The change is a result of a person's direct experiences as opposed to being in-born.
For the most part, the change an individual has made is permanent.
Effect on behavior
According to Bandura's social cognitive learning theory, observational learning can affect behavior in many ways, with both positive and negative consequences. It can teach completely new behaviors, for one. It can also increase or decrease the frequency of behaviors that have previously been learned. Observational learning can even encourage behaviors that were previously forbidden (for example, the violent behavior towards the Bobo doll that children imitated in Albert Bandura's study). Observational learning can also influence behaviors that are similar to, but not identical to, the ones being modeled. For example, seeing a model excel at playing the piano may motivate an observer to play the saxophone.
Age difference
Albert Bandura stressed that developing children learn from different social models, meaning that no two children are exposed to exactly the same modeling influence. From infancy to adolescence, they are exposed to various social models. A 2013 study found that a toddlers' previous social familiarity with a model was not always necessary for learning and that they were also able to learn from observing a stranger demonstrating or modeling a new action to another stranger.
It was once believed that babies could not imitate actions until the latter half of the first year. However, a number of studies now report that infants as young as seven days can imitate simple facial expressions. By the latter half of their first year, 9-month-old babies can imitate actions hours after they first see them. As they continue to develop, toddlers around age two can acquire important personal and social skills by imitating a social model.
Deferred imitation is an important developmental milestone in a two-year-old, in which children not only construct symbolic representations but can also remember information. Unlike toddlers, children of elementary school age are less likely to rely on imagination to represent an experience. Instead, they can verbally describe the model's behavior. Since this form of learning does not need reinforcement, it is more likely to occur regularly.
As age increases, age-related observational learning motor skills may decrease in athletes and golfers. Younger and skilled golfers have higher observational learning compared to older golfers and less skilled golfers.
Observational causal learning
Humans use observational Moleen causal learning to watch other people's actions and use the information gained to find out how something works and how we can do it ourselves.
A study of 25-month-old infants found that they can learn causal relations from observing human interventions. They also learn by observing normal actions not created by intentional human action.
Comparisons with imitation
Observational learning is presumed to have occurred when an organism copies an improbable action or action outcome that it has observed and the matching behavior cannot be explained by an alternative mechanism. Psychologists have been particularly interested in the form of observational learning known as imitation and in how to distinguish imitation from other processes. To successfully make this distinction, one must separate the degree to which behavioral similarity results from (a) predisposed behavior, (b) increased motivation resulting from the presence of another animal, (c) attention drawn to a place or object, (d) learning about the way the environment works, as distinguished from what we think of as (e) imitation (the copying of the demonstrated behavior).
Observational learning differs from imitative learning in that it does not require a duplication of the behavior exhibited by the model. For example, the learner may observe an unwanted behavior and the subsequent consequences, and thus learn to refrain from that behavior. For example, Riopelle (1960) found that monkeys did better with observational learning if they saw the "tutor" monkey make a mistake before making the right choice. Heyes (1993) distinguished imitation and non-imitative social learning in the following way: imitation occurs when animals learn about behavior from observing conspecifics, whereas non-imitative social learning occurs when animals learn about the environment from observing others.
Not all imitation and learning through observing is the same, and they often differ in the degree to which they take on an active or passive form. John Dewey describes an important distinction between two different forms of imitation: imitation as an end in itself and imitation with a purpose. Imitation as an end is more akin to mimicry, in which a person copies another's act to repeat that action again. This kind of imitation is often observed in animals. Imitation with a purpose utilizes the imitative act as a means to accomplish something more significant. Whereas the more passive form of imitation as an end has been documented in some European American communities, the other kind of more active, purposeful imitation has been documented in other communities around the world.
Observation may take on a more active form in children's learning in multiple Indigenous American communities. Ethnographic anthropological studies in Yucatec Mayan and Quechua Peruvian communities provide evidence that the home or community-centered economic systems of these cultures allow children to witness first-hand, activities that are meaningful to their own livelihoods and the overall well-being of the community. These children have the opportunity to observe activities that are relevant within the context of that community, which gives them a reason to sharpen their attention to the practical knowledge they are exposed to. This does not mean that they have to observe the activities even though they are present. The children often make an active decision to stay in attendance while a community activity is taking place to observe and learn. This decision underscores the significance of this learning style in many indigenous American communities. It goes far beyond learning mundane tasks through rote imitation; it is central to children's gradual transformation into informed members of their communities' unique practices. There was also a study, done with children, that concluded that Imitated behavior can be recalled and used in another situation or the same.
Apprenticeship
Apprenticeship can involve both observational learning and modelling. Apprentices gain their skills in part through working with masters in their profession and through observing and evaluating the work of their fellow apprentices. Examples include renaissance inventor/painter Leonardo da Vinci and Michelangelo, before succeeding in their profession they were apprentices.
Learning without imitation
Michael Tomasello described various ways of observational learning without the process of imitation in animals (ethology):
Exposure – Individuals learn about their environment through close proximity to other individuals that have more experience. For example, a young dolphin learning the location of a plethora of fish by staying near its mother.
Stimulus enhancement – Individuals become interested in an object from watching others interact with it. Increased interest in an object may result in object manipulation, which facilitates new object-related behaviors by trial-and-error learning. For example, a young killer whale might become interested in playing with a sea lion pup after watching other whales toss the sea lion pup around. After playing with the pup, the killer whale may develop foraging behaviors appropriate to such prey. In this case, the killer whale did not learn to prey on sea lions by observing other whales do so, but rather the killer whale became intrigued after observing other whales play with the pup. After the killer whale became interested, then its interactions with the sea lion resulted in behaviors that provoked future foraging efforts.
Goal emulation – Individuals are enticed by the end result of an observed behavior and attempt the same outcome but with a different method. For example, Haggerty (1909) devised an experiment in which a monkey climbed up the side of a cage, stuck its arm into a wooden chute, and pulled a rope in the chute to release food. Another monkey was provided an opportunity to obtain the food after watching a monkey go through this process on four separate occasions. The monkey performed a different method and finally succeeded after trial and error.
Peer model influences
Observational learning is very beneficial when there are positive, reinforcing peer models involved. Although individuals go through four different stages for observational learning: attention; retention; production; and motivation, this does not simply mean that when an individual's attention is captured that it automatically sets the process in that exact order. One of the most important ongoing stages for observational learning, especially among children, is motivation and positive reinforcement.
Performance is enhanced when children are positively instructed on how they can improve a situation and where children actively participate alongside a more skilled person. Examples of this are scaffolding and guided participation. Scaffolding refers to an expert responding contingently to a novice so the novice gradually increases their understanding of a problem. Guided participation refers to an expert actively engaging in a situation with a novice so the novice participates with or observes the adult to understand how to resolve a problem.
Cultural Variation
Cultural variation can be seen by the extent of information learned or absorbed by children in non-Western cultures through learning by observation. Cultural variation is not restricted only to ethnicity and nationality, but rather, extends to the specific practices within communities. In learning by observation, children use observation to learn without verbal requests for further information, or without direct instruction. For example, children from Mexican heritage families tend to learn and make better use of information observed during classroom demonstration than children of European heritage. Children of European heritage experience the type of learning that separates them from their family and community activities. They instead participate in lessons and other exercises in special settings such as school. Cultural backgrounds differ from each other in which children display certain characteristics in regards to learning an activity. Another example is seen in the immersion of children in some Indigenous communities of the Americas into the adult world and the effects it has on observational learning and the ability to complete multiple tasks simultaneously. This might be due to children in these communities having the opportunity to see a task being completed by their elders or peers and then trying to emulate the task. In doing so they learn to value observation and the skill-building it affords them because of the value it holds within their community. This type of observation is not passive, but reflects the child's intent to participate or learn within a community.
Observational learning can be seen taking place in many domains of Indigenous communities. The classroom setting is one significant example, and it functions differently for Indigenous communities compared to what is commonly present in Western schooling. The emphasis of keen observation in favor of supporting participation in ongoing activities strives to aid children to learn the important tools and ways of their community. Engaging in shared endeavors – with both the experienced and inexperienced – allows for the experienced to understand what the inexperienced need in order to grow in regards to the assessment of observational learning. The involvement of the inexperienced, or the children in this matter, can either be furthered by the children's learning or advancing into the activity performed by the assessment of observational learning. Indigenous communities rely on observational learning as a way for their children to be a part of ongoing activities in the community (Tharp, 2006).
Although learning in the Indigenous American communities is not always the central focus when participating in an activity, studies have shown that attention in intentional observation differs from accidental observation. Intentional participation is “keen observation and listening in anticipation of, or in the process of engaging in endeavors”. This means that when they have the intention of participating in an event, their attention is more focused on the details, compared to when they are accidentally observing.
Observational learning can be an active process in many Indigenous American communities. The learner must take initiative to attend to activities going on around them. Children in these communities also take initiative to contribute their knowledge in ways that will benefit their community. For example, in many Indigenous American cultures, children perform household chores without being instructed to do so by adults. Instead, they observe a need for their contributions, understand their role in their community, and take initiative to accomplish the tasks they have observed others doing. The learner's intrinsic motivations play an important role in the child's understanding and construction of meaning in these educational experiences. The independence and responsibility associated with observational learning in many Indigenous American communities are significant reasons why this method of learning involves more than just watching and imitating. A learner must be actively engaged with their demonstrations and experiences in order to fully comprehend and apply the knowledge they obtain.
Indigenous communities of the Americas
Children from indigenous heritage communities of the Americas often learn through observation, a strategy that can carry over into adulthood. The heightened value towards observation allows children to multi-task and actively engage in simultaneous activities. The exposure to an uncensored adult lifestyle allows children to observe and learn the skills and practices that are valued in their communities. Children observe elders, parents, and siblings complete tasks and learn to participate in them. They are seen as contributors and learn to observe multiple tasks being completed at once and can learn to complete a task while still engaging with other community members without being distracted.
Indigenous communities provide more opportunities to incorporate children in everyday life. This can be seen in some Mayan communities where children are given full access to community events, which allows observational learning to occur more often. Other children in Mazahua, Mexico are known to observe ongoing activities intensely . In native northern Canadian and indigenous Mayan communities, children often learn as third-party observers from stories and conversations by others. Most young Mayan children are carried on their mother's back, allowing them to observe their mother's work and see the world as their mother sees it. Often, children in Indigenous American communities assume the majority of the responsibility for their learning. Additionally, children find their own approaches to learning. Children are often allowed to learn without restrictions and with minimal guidance. They are encouraged to participate in the community even if they do not know how to do the work. They are self-motivated to learn and finish their chores. These children act as a second set of eyes and ears for their parents, updating them about the community.
Children aged 6 to 8 in an indigenous heritage community in Guadalajara, Mexico participated in hard work, such as cooking or running errands, thus benefiting the whole family, while those in the city of Guadalajara rarely did so. These children participated more in adult regulated activities and had little time to play, while those from the indigenous-heritage community had more time to play and initiate in their after-school activities and had a higher sense of belonging to their community. Children from formerly indigenous communities are more likely to show these aspects than children from cosmopolitan communities are, even after leaving their childhood community
Within certain indigenous communities, people do not typically seek out explanations beyond basic observation. This is because they are competent in learning through astute observation and often nonverbally encourage to do so. In a Guatemalan footloom factory, amateur adult weavers observed skilled weavers over the course of weeks without questioning or being given explanations; the amateur weaver moved at their own pace and began when they felt confident. The framework of learning how to weave through observation can serve as a model that groups within a society use as a reference to guide their actions in particular domains of life. Communities that participate in observational learning promote tolerance and mutual understand of those coming from different cultural backgrounds.
Other human and animal behavior experiments
When an animal is given a task to complete, they are almost always more successful after observing another animal doing the same task before them. Experiments have been conducted on several different species with the same effect: animals can learn behaviors from peers. However, there is a need to distinguish the propagation of behavior and the stability of behavior. Research has shown that social learning can spread a behavior, but there are more factors regarding how a behavior carries across generations of an animal culture.
Learning in fish
Experiments with ninespine sticklebacks showed that individuals will use social learning to locate food.
Social learning in pigeons
A study in 1996 at the University of Kentucky used a foraging device to test social learning in pigeons. A pigeon could access the food reward by either pecking at a treadle or stepping on it. Significant correspondence was found between the methods of how the observers accessed their food and the methods the initial model used in accessing the food.
Acquiring foraging niches
Studies have been conducted at the University of Oslo and University of Saskatchewan regarding the possibility of social learning in birds, delineating the difference between cultural and genetic acquisition. Strong evidence already exists for mate choice, bird song, predator recognition, and foraging.
Researchers cross-fostered eggs between nests of blue tits and great tits and observed the resulting behavior through audio-visual recording. Tits raised in the foster family learned their foster family's foraging sites early. This shift—from the sites the tits would among their own kind and the sites they learned from the foster parents—lasted for life. What young birds learn from foster parents, they eventually transmitted to their own offspring. This suggests cultural transmissions of foraging behavior over generations in the wild.
Social learning in crows
The University of Washington studied this phenomenon with crows, acknowledging the evolutionary tradeoff between acquiring costly information firsthand and learning that information socially with less cost to the individual but at the risk of inaccuracy. The experimenters exposed wild crows to a unique “dangerous face” mask as they trapped, banded, and released 7-15 birds at five different study places around Seattle, WA. An immediate scolding response to the mask after trapping by previously captured crows illustrates that the individual crow learned the danger of that mask. There was a scolding from crows that were captured that had not been captured initially. That response indicates conditioning from the mob of birds that assembled during the capture.
Horizontal social learning (learning from peers) is consistent with the lone crows that recognized the dangerous face without ever being captured. Children of captured crow parents were conditioned to scold the dangerous mask, which demonstrates vertical social learning (learning from parents). The crows that were captured directly had the most precise discrimination between dangerous and neutral masks than the crows that learned from the experience of their peers. The ability of crows to learn doubled the frequency of scolding, which spread at least 1.2 km from where the experiment started to over a 5-year period at one site.
Propagation of animal culture
Researchers at the Département d’Etudes Cognitives, Institut Jean Nicod, Ecole Normale Supérieure acknowledged a difficulty with research in social learning. To count acquired behavior as cultural, two conditions need must be met: the behavior must spread in a social group, and that behavior must be stable across generations. Research has provided evidence that imitation may play a role in the propagation of a behavior, but these researchers believe the fidelity of this evidence is not sufficient to prove the stability of animal culture.
Other factors like ecological availability, reward-based factors, content-based factors, and source-based factors might explain the stability of animal culture in a wild rather than just imitation. As an example of ecological availability, chimps may learn how to fish for ants with a stick from their peers, but that behavior is also influenced by the particular type of ants as well as the condition. A behavior may be learned socially, but the fact that it was learned socially does not necessarily mean it will last. The fact that the behavior is rewarding has a role in cultural stability as well. The ability for socially-learned behaviors to stabilize across generations is also mitigated by the complexity of the behavior. Different individuals of a species, like crows, vary in their ability to use a complex tool. Finally, a behavior's stability in animal culture depends on the context in which they learn a behavior. If a behavior has already been adopted by a majority, then the behavior is more likely to carry across generations out of a need for conforming.
Animals are able to acquire behaviors from social learning, but whether or not that behavior carries across generations requires more investigation.
Hummingbird experiment
Experiments with hummingbirds provided one example of apparent observational learning in a non-human organism. Hummingbirds were divided into two groups. Birds in one group were exposed to the feeding of a knowledgeable "tutor" bird; hummingbirds in the other group did not have this exposure. In subsequent tests the birds that had seen a tutor were more efficient feeders than the others.
Bottlenose dolphin
Herman (2002) suggested that bottlenose dolphins produce goal-emulated behaviors rather than imitative ones. A dolphin that watches a model place a ball in a basket might place the ball in the basket when asked to mimic the behavior, but it may do so in a different manner seen.
Rhesus monkey
Kinnaman (1902) reported that one rhesus monkey learned to pull a plug from a box with its teeth to obtain food after watching another monkey succeed at this task.
Fredman (2012) also performed an experiment on observational behavior. In experiment 1, human-raised monkeys observed a familiar human model open a foraging box using a tool in one of two alternate ways: levering or poking. In experiment 2, mother-raised monkeys viewed similar techniques demonstrated by monkey models. A control group in each population saw no model. In both experiments, independent coders detected which technique experimental subjects had seen, thus confirming social learning. Further analyses examined copying at three levels of resolution.
The human-raised monkeys exhibited the greatest learning with the specific tool use technique they saw. Only monkeys who saw the levering model used the lever technique, by contrast with controls and those who witnessed poking. Mother-reared monkeys instead typically ignored the tool and exhibited fidelity at a lower level, tending only to re-create whichever result the model had achieved by either levering or poking.
Nevertheless, this level of social learning was associated with significantly greater levels of success in monkeys witnessing a model than in controls, an effect absent in the human-reared population. Results in both populations are consistent with a process of canalization of the repertoire in the direction of the approach witnessed, producing a narrower, socially shaped behavioral profile than among controls who saw no model.
Light box experiment
Pinkham and Jaswal (2011) did an experiment to see if a child would learn how to turn on a light box by watching a parent. They found that children who saw a parent use their head to turn on the light box tended to do the task in that manner, while children who had not seen the parent used their hands instead.
Swimming skill performance
When adequate practice and appropriate feedback follow demonstrations, increased skill performance and learning occurs. Lewis (1974) did a study of children who had a fear of swimming and observed how modelling and going over swimming practices affected their overall performance. The experiment spanned nine days, and included many steps. The children were first assessed on their anxiety and swimming skills. Then they were placed into one of three conditional groups and exposed to these conditions over a few days.
At the end of each day, all children participated in a group lesson. The first group was a control group where the children watched a short cartoon video unrelated to swimming. The second group was a peer mastery group, which watched a short video of similar-aged children who had very good task performances and high confidence. Lastly, the third group was a peer coping group, whose subjects watched a video of similar-aged children who progressed from low task performances and low confidence statements to high task performances and high confidence statements.
The day following the exposures to each condition, the children were reassessed. Finally, the children were also assessed a few days later for a follow up assessment. Upon reassessment, it was shown that the two model groups who watched videos of children similar in age had successful rates on the skills assessed because they perceived the models as informational and motivational.
Do-as-I-do Chimpanzee
Flexible methods must be used to assess whether an animal can imitate an action. This led to an approach that teaches animals to imitate by using a command such as “do-as-I-do" or “do this” followed by the action that they are supposed to imitate . Researchers trained chimpanzees to imitate an action that was paired with the command. For example, this might include a researcher saying “do this” paired with clapping hands. This type of instruction has been utilized in a variety of other animals in order to teach imitation actions by utilizing a command or request.
Observational learning in Everyday Life
Observational learning allows for new skills to be learned in a wide variety of areas. Demonstrations help the modification of skills and behaviors.
Learning physical activities
When learning skills for physical activities can be anything that is learned that requires physical movement, this can include learning a sport, learning to eat with a fork, or learning to walk. There are multiple important variables that aid in modifying physical skills and psychological responses from an observational learning standpoint. Modeling is a variable in observational learning where the skill level of the model is considered. When someone is supposed to demonstrate a physical skill such as throwing a baseball the model should be able to execute the behavior of throwing the ball flawlessly if the model of learning is a mastery model. Another model to utilize in observational learning is a coping model, which would be a model demonstrating a physical skill that they have not yet mastered or achieved high performance in. Both models are found to be effective and can be utilized depending on the what skills is trying to be demonstrated. These models can be used as interventions to increase observational learning in practice, competition, and rehabilitation situations.
Neuroscience
Recent research in neuroscience has implicated mirror neurons as a neurophysiological basis for observational learning. These specialized visuomotor neurons fire action potentials when an individual performs a motor task and also fire when an individual passively observes another individual performing the same motor task. In observational motor learning, the process begins with a visual presentation of another individual performing a motor task, this acts as a model. The learner then needs to transform the observed visual information into internal motor commands that will allow them to perform the motor task, this is known as visuomotor transformation. Mirror neuron networks provide a mechanism for visuo-motor and motor-visual transformation and interaction. Similar networks of mirror neurons have also been implicated in social learning, motor cognition and social cognition.
Clinical Perspective
Autism Spectrum Disorder
Discreet trial training (DTT) is a structured and systematic approach utilized in helping individuals with autism spectrum disorder learn. Individuals with autism tend to struggle with learning through observation, therefore something that is reinforcing is necessary in order to motivate them to imitate or follow through with the task. When utilizing DTT to teach individuals with autism modeling is utilized to aid in their learning. Modeling would include showing how to reach the correct answer, this could mean showing the steps to a math equation. Utilizing DTT in a group setting also promotes observational learning from peers as well.
See also
Cognitive imitation
Educational psychology
Educational technology
Hypercorrection
Imitation
Inference
Machine learning
Mirroring (psychology)
Social learning theory
Social learning tools
Mathematical models of social learning
References
Further reading on animal social learning
Zentall, T.R. (2006). Imitation: Definitions, evidence, and mechanisms. Animal Cognition, 9 335–353. (A thorough review of different types of social learning) Full text
Social learning theory
Behavioral concepts
Learning
Animal cognition |
null | null | History of Russia (1991–present) | eng_Latn | The history of Russia from 1991 to the present began with the dissolution of the Soviet Union (USSR) on 26 December 1991. The Russian Soviet Federative Socialist Republic (RSFSR) was the largest republic within the USSR, but until 1990 it had no significant independence.
The RSFSR was the largest of the fifteen republics that made up the USSR, accounting for over 60% of the gross domestic product (GDP) and over 50% of the Soviet population. Russians also dominated the Soviet military and the Communist Party (CPSU). As such, the Russian Federation was widely accepted as the USSR's successor state in diplomatic affairs and it assumed the USSR's permanent membership and veto in the UN Security Council (see Russia and the United Nations).
Prior to the dissolution of the USSR, Boris Yeltsin had been elected President of Russia in June 1991 in the first direct presidential election in Russian history. This ensured that Yeltsin would be the political leader of the Russian successor state following dissolution. This situation resulted in political turmoil as the Soviet and Russian leadership wrestled for control, which culminated in the 1991 August coup, where the Soviet military attempted to overthrow Mikhail Gorbachev. Although the coup was ultimately averted, this situation contributed to rising instability in the Soviet Union. As the USSR was on the verge of collapse by October 1991, Yeltsin announced that Russia would proceed with radical reforms, including market-oriented reform along the lines of Poland's "big bang", also known as "shock therapy". Following Yeltsin's resignation in 1999, Russia's politics have since been dominated by Vladimir Putin, serving as either President or Prime Minister. Although the Russian economy has improved significantly under Putin's leadership following relative economic chaos under Yeltsin, Putin has also been widely accused of corruption, authoritarian leadership, and widespread human rights abuses.
For the most part, the Russian armed forces were in near complete disarray by 1992, one year after dissolution. This degraded military effectiveness would become all too clear during the 1994 Chechen War, but in the interim this posed some significant practical challenges for global security and arms control. Under Russian leadership, the Lisbon Protocol ensured that former Soviet republics would disarm themselves of nuclear weapons. This may have been particularly important for Kazakhstan which hosted a significant share of the world's nuclear weapons immediately following the dissolution of the Soviet Union. However, the former Soviet republics were able to maintain transnational cooperation in other military areas, like establishing shared responsibility for the rocket and space infrastructure such as the Baikonur Cosmodrome.
Reforms
"Shock therapy"
With the dissolution of the Warsaw Pact and CoMEcon and other treaties that served to bind its satellite states to the Soviet Union, the conversion of the world's largest state-controlled economy into a market-oriented economy would have been extraordinarily difficult regardless of the policies chosen. The policies chosen for this difficult transition were (1) liberalization, (2) stabilization and (3) privatization. These policies were based on the neoliberal Washington Consensus of the International Monetary Fund (IMF), World Bank and Treasury Department.
The programs of liberalization and stabilization were designed by Yeltsin's deputy prime minister Yegor Gaidar, a 35-year-old liberal economist inclined toward radical reform, and widely known as an advocate of "shock therapy". Shock therapy was originally used in Bolivia by notable economist Jeffery Sachs to combat inflation in the 1980s. Having achieved some major successes in Bolivia, shock therapy was then imported to the Polish context following the dissolution of the Soviet Union, and Russia shortly after.
The partial results of liberalization (lifting price controls) included worsening already apparent hyperinflation, initially due to monetary overhang and exacerbated after the central bank, an organ under parliament, which was skeptical of Yeltsin's reforms, was short of revenue and printed money to finance its debt. This resulted in the near bankruptcy of much of Russian industry.
The process of liberalization would create winners and losers, depending on how particular industries, classes, age groups, ethnic groups, regions, and other sectors of Russian society were positioned. Some would benefit by the opening of competition; others would suffer. Among the winners were the new class of entrepreneurs and black marketeers that had emerged under Mikhail Gorbachev's perestroika. But liberalizing prices meant that the elderly and others on fixed incomes would suffer a severe drop in living standards, and people would see a lifetime of savings wiped out.
With inflation at double-digit rates per month as a result of printing, macroeconomic stabilization was enacted to curb this trend. Stabilization, also called structural adjustment, is a harsh austerity regime (tight monetary policy and fiscal policy) for the economy in which the government seeks to control inflation. Under the stabilization program, the government let most prices float, raised interest rates to record highs, raised heavy new taxes, sharply cut back on government subsidies to industry and construction, and made massive cuts in state welfare spending. These policies caused widespread hardship as many state enterprises found themselves without orders or financing. A deep credit crunch shut down many industries and brought about a protracted depression.
The rationale of the program was to squeeze the built-in inflationary pressure out of the economy so that producers would begin making sensible decisions about production, pricing and investment instead of chronically overusing resources—a problem that resulted in shortages of consumer goods in the Soviet Union in the 1980s. By letting the market rather than central planners determine prices, product mixes, output levels, and the like, the reformers intended to create an incentive structure in the economy where efficiency and risk would be rewarded and waste and carelessness were punished. Removing the causes of chronic inflation, the reform architects argued, was a precondition for all other reforms: Hyperinflation would wreck both democracy and economic progress, they argued; they also argued that only by stabilizing the state budget could the government proceed to dismantle the Soviet planned economy and create a new capitalist Russia.
Obstacles to reform
The former Soviet Union had to deal with a number of unique obstacles during the post-Soviet transition including political reform, economic restructuring and the redrawing of political boundaries. The discomfort associated with these changes was not felt the same in each former Soviet republic. As a general rule, states to Russia's west, such as Poland, Hungary, and the Czech Republic, have fared slightly better than their eastern neighbors since the collapse of the Eastern bloc, while Russia itself and countries to Russia's east experienced greater difficulties and found themselves on worse footing immediately after dissolution. A major reason that Russia's transition has been so wrenching is that it is remaking both its Soviet-era political and economic institutions at once. In addition to institutional reforms designed to create a new political-economic system, Russia was also charged with remaking itself into a new national state following the disintegration of the Soviet Union.
The first major problem facing Russia was the legacy of the Soviet Union's enormous commitment to the Cold War. In the late 1980s, the Soviet Union devoted a quarter of its gross economic output to the defense sector (at the time most Western analysts believed that this figure was 15 percent). At the time, the military-industrial complex employed at least one of every five adults in the Soviet Union. In some regions of Russia, at least half of the workforce was employed in defense plants (the comparable U.S. figures were roughly one-sixteenth of gross national product and about one of every sixteen in the workforce). These over-reliance on the military sector made Russian industry and human capital relatively noncompetitive upon entry into a market-oriented system. Furthermore, the end of the Cold War and the cutback in military spending affected industry quite dramatically making it difficult to quickly retool equipment, retrain workers, and find new markets. In the process of economic re-tooling, an enormous body of experience, qualified specialists and know-how was lost or misallocated, as the plants were sometimes switching from, for example, producing hi-tech military equipment to making kitchen utensils.
A second obstacle, partly related to the sheer vastness and geographical diversity of the Russian landmass, was the sizable number of "mono-industrial" regional economies (regions dominated by a single industrial employer) that Russia inherited from the Soviet Union. The concentration of production in a relatively small number of big state enterprises meant that many local governments were entirely dependent on the economic health of a single employer; when the Soviet Union collapsed and the economic ties between Soviet republics and even regions were severed, the production in the whole country dropped by more than fifty percent. Roughly half of Russia's cities had only one large industrial enterprise, and three fourths had no more than four. Consequently, the decrease in production caused tremendous unemployment and underemployment.
Thirdly, post-Soviet Russia did not inherit a system of state social security and welfare from the USSR. Instead the companies, mainly large industrial firms, were traditionally responsible for a broad range of social welfare functions—building and maintaining housing for their work forces, and managing health, recreational, educational, and similar facilities. The towns in contrast possessed neither the apparatus nor the funds for the provision of basic social services. Industrial employees were left heavily dependent on their firms. Thus, economic transformation created severe problems in maintaining social welfare since local governments were unable to assume finance or operational responsibility for these functions.
Finally, there is a human capital dimension to the failure of post-Soviet reforms in Russia. The former Soviet population was not necessarily uneducated. Literacy was nearly universal, and the educational level of the Soviet population was among the highest in the world with respect to science, engineering, and some technical disciplines, although the Soviets devoted little to what would be described as "liberal arts" in the West. With the move to a post-Communist system, the Russian university system collapsed. Rampant credential inflation in the Russian university system made it difficult for employers to determine who was really skilled and the problems of the higher education system more generally made it difficult to remedy other issues of human capital that came from the transition to a market-oriented system, such as upskilling and re-skilling. For example, former state enterprise managers were highly skilled at coping with the demands on them under the Soviet system of planned production targets, but discouraged the risk-and-reward centered behavior of market capitalism. These managers were responsible for a broad array of social welfare functions for their employees, their families, and the population of the towns and regions where they were located. Profitability and efficiency, however, were generally not the most prominent priorities for Soviet enterprise managers. Thus, almost no Soviet employees or managers had firsthand experience with decision-making in the conditions of a market economy.
Depression
After the initial turmoil and euphoria of early marketizations, Russia's economy sank into deep depression by the mid-1990s due to botched reform efforts and low commodity prices globally but not before George H. W. Bush helped Yeltsin with "an unparalleled opportunity to change the nuclear posture of both the United States and the Soviet Union" and to end the Cold War peacefully with the Nunn–Lugar cash-for-weapons scheme, in order to avoid the worst of the dissolution of the vast Soviet technological empire. Russia's economy was hit further by the financial crash of 1998 before experiencing a modest recovery in 1999–2000 as commodity prices began to rise again. According to Russian government statistics, the economic decline was far more severe than the Great Depression was in the United States in terms of gross domestic product. By way of a domestic comparison, the post-Soviet economic decline was about half as severe as the economic catastrophe borne out of the immediate consequence of World War I, the fall of Tsarism, and the Russian Civil War.
Following the economic collapse of the early 1990s, Russia suffered from a sharp increase in the rates of poverty and economic inequality. Estimates by the World Bank based on both macroeconomic data and surveys of household incomes and expenditures indicate that whereas 1.5% of the population was living in poverty (defined as income below the equivalent of $25 per month) in the late Soviet era, by mid-1993 between 39% and 49% of the population was living in poverty. Per capita incomes fell by another 15% by 1998, according to government figures.
Public health indicators show a dramatic corresponding decline. Although all post-Soviet countries experience an immediate decline in birth-rates due to economic turmoil this may have been particularly acute in Russia. In 1999, total population fell by about three-quarters of a million people. Meanwhile, life expectancy dropped for men from 64 years in 1990 to 57 years by 1994, while women's dropped from 74 to about 71. Both health factors and a sharp increase in deaths of the youth demographic from unnatural causes (such as murders, suicides, and accidents) have significantly contributed to this trend. Closely related to the declining life expectancy, alcohol-related deaths skyrocketed 60% in the 1990s and deaths from infectious and parasitic diseases shot up 100%, mainly because medicines were no longer affordable to the poor.
While Russia no longer suffered from the supply shortages of consumer goods that were so characteristic of the 1980s USSR (see Consumer goods in the Soviet Union), this was not only related to the opening of the Russian market to imports in the early 1990s but also to the relative impoverishment of the Russian people in the 1980s. Russians on fixed incomes (the vast majority of the workforce) saw their purchasing power drastically reduced, so while the stores might have been well stocked in the Yeltsin era, average people could now afford to buy little, if anything from these stores. By 2011, the average income has risen to more than $700 per month, emblematic of the mild recovery in recent years thanks to a large extent to high oil prices. The growing income, however, has not been evenly distributed. The social inequality has risen sharply during the 1990s with the Gini coefficient, for example, reaching 42% by the end of 2010. Russia's income disparities are now nearly as large as Brazil (which has long been among the world leader in inequality) while regional disparities in the level of poverty continue to trend upwards.
Backlash
Structural reform and a severe devaluation of the ruble lowered the standard of living for most segments of the Russian population. As a result, there was powerful political opposition to reform. Democratization opened the political channels for venting these frustrations, which translated into votes for anti-reform candidates, especially those of the Communist Party of the Russian Federation and its allies in the Duma. Russian voters, able to vote for opposition parties in the 1990s, often rejected economic reforms and yearned for the stability and personal security of the Soviet era. These were the groups that had enjoyed the benefits of Soviet-era state-controlled wages and prices, high state spending to subsidize priority sectors of the economy, protection from competition with foreign industries, and welfare entitlement programs. During the Yeltsin years in the 1990s, these anti-reformist groups were well organized, voicing their opposition to reform through strong trade unions, associations of directors of state-owned firms, and political parties in the popularly elected parliament whose primary constituencies were among those vulnerable to reform. A constant theme of Russian history in the 1990s was the conflict between economic reformers and those hostile to the new capitalism.
Reform by decree
On January 2, 1992, Yeltsin—acting as his own prime minister—enacted the most comprehensive components of economic reform by decree, thereby circumventing the Supreme Soviet of Russia and Congress of People's Deputies of Russia, which had been elected in March 1990, before the dissolution of the USSR. While this spared Yeltsin from the prospects of bargaining and wrangling with Soviet deputies, it also eliminated any meaningful discussion of the right course of action for the country. Nonetheless, radical reform continued to face some critical political barriers. The Soviet-era Central Bank was still subordinate to the conservative Supreme Soviet who continued to support socialist policies in opposition to Yeltsin and the presidency. During the height of hyperinflation in 1992–1993, the Central Bank actually tried to derail reforms by actively printing even more money during this period of inflation. After all, the Russian government was short of revenue and was forced to print money to finance its debt. As a result, inflation exploded into hyperinflation, and the Russian economy continued into an evermore serious slump.
Crisis
Constitutional crisis
The struggle for the center of power in Soviet Russia following the collapse of the Soviet Union and for the nature of the economic reforms culminated in a political crisis and bloodshed in the autumn of 1993. Yeltsin, who represented a course of radical privatization, was opposed by the Supreme Soviet. Confronted with opposition to the presidential power of decree and threatened with impeachment, he "dissolved" the parliament on September 21, in contravention of the existing constitution, and ordered new elections and a referendum on a new constitution. The parliament then declared Yeltsin deposed and appointed Aleksandr Rutskoy acting president on September 22. Tensions built quickly, and matters came to a head after street riots on October 2–October 3. On October 4, Yeltsin ordered Special Forces and elite army units to storm the parliament building, the "White House" as it is called. With tanks thrown against the small-arms fire of the parliamentary defenders, the outcome was not in doubt. Rutskoy, Ruslan Khasbulatov, and the other parliamentary supporters surrendered and were immediately arrested and jailed. The official count was 147 dead, 437 wounded (with several men killed and wounded on the presidential side).
Thus the transitional period in post-Soviet Russian politics came to an end. A new constitution was approved by referendum in December 1993. Russia was given a strongly presidential system. Radical privatization went ahead. Although the old parliamentary leaders were released without trial on February 26, 1994, they would not play an open role in politics thereafter. Though its clashes with the executive would eventually resume, the remodeled Russian parliament had greatly circumscribed powers. (For details on the constitution passed in 1993 see Constitution and government structure of Russia.)
First Chechen War
In 1994, Yeltsin dispatched 40,000 troops to the southern region of Chechnya to prevent its secession from Russia. Living south of Moscow, the predominantly Muslim Chechens for centuries had gloried in defying the Russians. Dzhokhar Dudayev, the Republic of Chechnya's nationalist president, was driven to take his republic out of the Russian Federation, and had declared Chechnya's independence in 1991. Russia was quickly submerged in a quagmire like that of the U.S. in the Vietnam War. When the Russians attacked the Chechen capital of Grozny during the first weeks of January 1995, about 25,000 civilians died under week-long air raids and artillery fire in the sealed-off city. Massive use of artillery and air-strikes remained the dominating strategy throughout the Russian campaign. Even so, Chechen forces seized thousands of Russian hostages, while inflicting humiliating losses on the demoralized and ill-equipped Russian troops. Russian troops had not secured the Chechen capital of Grozny by year's end.
The Russians finally managed to gain control of Grozny in February 1995 after heavy fighting. In August 1996, Yeltsin agreed to a ceasefire with Chechen leaders, and a peace treaty was formally signed in May 1997. However, the conflict resumed in 1999, thus rendering the 1997 peace accord meaningless. This time the rebellion was crushed by Vladimir Putin.
Rise of the oligarchs
The new capitalist opportunities presented by the opening of the Russian economy in the late 1980s and early 1990s affected many people's interests. As the Soviet system was being dismantled, well-placed bosses and technocrats in the Communist Party, KGB, and Komsomol (Soviet Youth League) were cashing in on their Soviet-era power and privileges. Some quietly liquidated the assets of their organization and secreted the proceeds in overseas accounts and investments. Others created banks and business in Russia, taking advantage of their insider positions to win exclusive government contracts and licenses and to acquire financial credits and supplies at artificially low, state-subsidized prices in order to transact business at high, market-value prices. Great fortunes were made almost overnight.
At the same time, a few young people, without much social status, saw opportunity in the economic and legal confusion of the transition. Between 1987 and 1992, trading of natural resources and foreign currencies, as well as imports of highly demanded consumer goods and then domestic production of their rudimentary substitutes, rapidly enabled these pioneering entrepreneurs to accumulate considerable wealth. In turn, the emerging cash-based, highly opaque markets provided a breeding ground for a large number of racket gangs.
By the mid-1990s, the best-connected former nomenklatura leaders accumulated considerable financial resources, while on the other hand, the most successful entrepreneurs became acquainted with government officials and public politicians. The privatization of state enterprises was a unique opportunity because it gave many of those who had gained wealth in the early 1990s a chance to convert it into shares of privatized enterprises.
The Yeltsin government hoped to use privatization to spread ownership of shares in former state enterprises as widely as possible to create political support for his government and his reforms. The government used a system of free vouchers as a way to give mass privatization a jump-start. But it also allowed people to purchase shares of stock in privatized enterprises with cash. Even though initially each citizen received a voucher of equal face value, within months most of the vouchers converged in the hands of intermediaries who were ready to buy them for cash right away.
As the government ended the voucher privatization phase and launched cash privatization, it devised a program that it thought would simultaneously speed up privatization and yield the government a much-needed infusion of cash for its operating needs. Under the scheme, which quickly became known in the West as "loans for shares," the Yeltsin regime auctioned off substantial packages of stock shares in some of its most desirable enterprises, such as energy, telecommunications, and metallurgical firms, as collateral for bank loans.
In exchange for the loans, the state handed over assets worth many times as much. Under the terms of the deals, if the Yeltsin government did not repay the loans by September 1996, the lender acquired title to the stock and could then resell it or take an equity position in the enterprise. The first auctions were held in the fall of 1995. The auctions themselves were usually held in such a way so to limit the number of banks bidding for shares and thus to keep the auction prices extremely low. By summer 1996, major packages of shares in some of Russia's largest firms had been transferred to a small number of major banks, thus allowing a handful of powerful banks to acquire substantial ownership shares over major firms at shockingly low prices. These deals were effectively giveaways of valuable state assets to a few powerful, well-connected, and wealthy financial groups.
The concentration of immense financial and industrial power, which loans for shares had assisted, extended to the mass media. One of the most prominent of the financial barons Boris Berezovsky, who controlled major stakes in several banks and companies, exerted an extensive influence over state television programming for a while. Berezovsky and other ultra-wealthy, well-connected tycoons who controlled these great empires of finance, industry, energy, telecommunications, and media became known as the "Russian oligarchs". Along with Berezovsky, Mikhail Khodorkovsky, Roman Abramovich, Vladimir Potanin, Vladimir Bogdanov, Rem Viakhirev, Vagit Alekperov, Viktor Chernomyrdin, Viktor Vekselberg, and Mikhail Fridman emerged as Russia's most powerful and prominent oligarchs.
A tiny clique who used their connections built up during the last days of the Soviet years to appropriate Russia's vast resources during the rampant privatizations of the Yeltsin years, the oligarchs emerged as the most hated men in the nation. The Western world generally advocated a quick dismantling of the Soviet planned economy to make way for "free-market reforms," but later expressed disappointment over the newfound power and corruption of the "oligarchs."
Presidential election of 1996
Campaigns
Early in the campaign it had been thought that Yeltsin, who was in uncertain health (after recuperating from a series of heart attacks) and whose behavior was sometimes erratic, had little chance for reelection. When campaigning opened at the beginning of 1996, Yeltsin's popularity was close to zero. Meanwhile, the opposition Communist Party of the Russian Federation had already gained ground in parliamentary voting on December 17, 1995, and its candidate, Gennady Zyuganov, had a strong grassroots organization, especially in the rural areas and small towns, and appealed effectively to memories of the old days of Soviet prestige on the international stage and the socialist domestic order.
Panic struck the Yeltsin team when opinion polls suggested that the ailing president could not win; members of his entourage urged him to cancel presidential elections and effectively rule as dictator from then on. Instead, Yeltsin changed his campaign team, assigning a key role to his daughter, Tatyana Dyachenko, and appointing Anatoly Chubais campaign manager. Chubais, who was not just Yeltsin's campaign manager but also the architect of Russia's privatization program, set out to use his control of the privatization program as the key instrument of Yeltsin's reelection campaign.
The president's inner circle assumed that it had only a short time in which to act on privatization; it, therefore, needed to take steps that would have a large and immediate impact, making the reversal of reform prohibitively costly for their opponents. Chubais' solution was to co-opt potentially powerful interests, including enterprise directors and regional officials, in order to ensure Yeltsin's reelection.
The position of the enterprise directors to the program was essential to maintaining economic and social stability in the country. The managers represented one of the most powerful collective interests in the country; it was the enterprise managers who could ensure that labor did not erupt in a massive wave of strikes. The government, therefore, did not strenuously resist the tendency for voucher privatization to turn into "insider privatization," as it was termed, in which senior enterprise officials acquired the largest proportion of shares in privatized firms. Thus, Chubais allowed well-connected employees to acquire majority stakes in the enterprises. This proved to be the most widely used form of privatization in Russia. Three-quarters of privatized enterprises opted for this method, most often using vouchers. Real control thus wound up in the hands of the managers.
Support from the oligarchs was also crucial to Yeltsin's reelection campaign. The "loans for shares" giveaway took place in the run-up to the 1996 presidential election—at a point when it had appeared that Zyuganov might defeat Yeltsin. Yeltsin and his entourage gave the oligarchs an opportunity to scoop up some of Russia's most desirable assets in return for their help in his reelection effort. The oligarchs, in turn, reciprocated the favor.
In the spring of 1996, with Yeltsin's popularity at a low ebb, Chubais and Yeltsin recruited a team of six leading Russian financiers and media barons (all oligarchs) who bankrolled the Yeltsin campaign with $3 million and guaranteed coverage on television and in leading newspapers directly serving the president's campaign strategy. The media painted a picture of a fateful choice for Russia, between Yeltsin and a "return to totalitarianism." The oligarchs even played up the threat of civil war if a Communist were elected president.
In the outlying regions of the country, the Yeltsin campaign relied on its ties to other allies—the patron-client ties of the local governors, most of whom had been appointed by the president.
The Zyuganov campaign had a strong grass-roots organization, but it was simply no match for the financial resources and access to patronage that the Yeltsin campaign could marshal.
Yeltsin campaigned energetically, dispelling concerns about his health, exploiting all the advantages of incumbency to maintain a high media profile. To assuage voters' discontent, he made the claim that he would abandon some unpopular economic reforms and boost welfare spending, end the war in Chechnya, pay wage and pension arrears, and abolish military conscription (he did not live up to his promises after the election, except for ending the Chechen war, which was halted for 3 years). Yeltsin's campaign also got a boost from the announcement of a $10 billion loan to the Russian government from the International Monetary Fund.
Grigory Yavlinsky was the liberal alternative to Yeltsin and Zyuganov. He appealed to a well-educated middle class that saw Yeltsin as an incompetent alcoholic and Zyuganov as a Soviet-era throwback. Seeing Yavlinsky as a threat, Yeltsin's inner circle of supporters worked to bifurcate political discourse, thus excluding a middle ground—and convince voters that only Yeltsin could defeat the Communist "menace." The election became a two-man race, and Zyuganov, who lacked Yeltsin's resources and financial backing, watched helplessly as his strong initial lead was whittled away.
Elections
Voter turnout in the first round of the polling on June 16 was 69.8%. According to returns announced on June 17, Yeltsin won 35% of the vote; Zyuganov won 32%; Aleksandr Lebed, a populist ex-general, a surprisingly high 14.5%; liberal candidate Grigory Yavlinsky 7.4%; far-right nationalist Vladimir Zhirinovsky 5.8%; and former Soviet president Mikhail Gorbachev 0.5%. With no candidate securing an absolute majority, Yeltsin and Zyuganov went into a second round of voting. In the meantime, Yeltsin co-opted a large segment of the electorate by appointing Lebed to the posts of national security adviser and secretary of the Security Council.
In the end, Yeltsin's election tactics paid off. In the run-off on July 3, with a turnout of 68.9%, Yeltsin won 53.8% of the vote and Zyuganov 40.3%, with the rest (5.9%) voting "against all". Moscow and Saint Petersburg (formerly Leningrad) together provided over half of the incumbent president's support, but he also did well in large cities in the Urals and in the north and northeast. Yeltsin lost to Zyuganov in Russia's southern industrial heartland. The southern stretch of the country became known as the "red belt", underscoring the resilience of the Communist Party in elections since the breakup of the Soviet Union.
Although Yeltsin promised that he would abandon his unpopular neoliberal austerity policies and increase public spending to help those suffering from the pain of capitalist reforms, within a month of his election, Yeltsin issued a decree canceling almost all of these promises.
Right after the election, Yeltsin's physical health and mental stability were increasingly precarious. Many of his executive functions thus devolved upon a group of advisers (most of whom had close links with the oligarchs).
Financial collapse
The global recession of 1998, which started with the Asian financial crisis in July 1997, exacerbated Russia's continuing economic crisis. Given the ensuing decline in world commodity prices, countries heavily dependent on the export of raw materials such as oil were among those most severely hit. Oil, natural gas, metals, and timber account for more than 80% of Russian exports, leaving the country vulnerable to swings in world prices. Oil is also a major source of government tax revenue which brought significant negative implications for Russia's fiscal situation, foreign exchange stores and ultimately, the value of the ruble.
The pressures on the ruble, reflecting the weakness of the economy, resulted in a disastrous fall in the value of the currency. Massive tax evasion continued and accelerated due to financial instability and decreasing government capacity. This further decreased government revenues and soon, the central government found itself unable to service the massive loans it had accumulated and ultimately was even unable to pay its employees. The government stopped making timely payment of wages, pensions, and debts to suppliers; and when workers were paid, it was often with bartered goods rather than rubles. Coal miners were especially hard hit, and for several weeks in the summer they blocked sections of the Trans-Siberian railroad with protests, effectively cutting the country in two. As time wore on, they added calls for the resignation of Yeltsin in addition to their demands for wages.
A political crisis came to a head in March when Yeltsin suddenly dismissed Prime Minister Viktor Chernomyrdin and his entire cabinet on March 23. Yeltsin named a virtually unknown technocrat, Energy Minister Sergei Kiriyenko, aged 35, as acting prime minister. Russian observers expressed doubts about Kiriyenko's youth and inexperience. The Duma rejected his nomination twice. Only after a month-long standoff, during which Yeltsin threatened to dissolve the legislature, did the Duma confirm Kiriyenko on a third vote on April 24.
Kiriyenko appointed a new cabinet strongly committed to stemming the fall in the value of Russia's currency. The oligarchs strongly supported Kiriyenko's efforts to maintain the exchange rate. A high exchange rate meant that they needed fewer rubles to buy imported goods, especially luxury items.
In an effort to prop up the currency and stem the flight of capital, Kiriyenko hiked interest rates to 150% in order to attract buyers for government bonds. But concerns about the financial crisis in Asia and the slump in world oil prices were already prompting investors to withdraw from Russia. By mid-1998, it was clear Russia would need help from IMF to maintain its exchange rate.
The Russian crisis caused alarm in the West. Pouring more money into the Russian economy would not be a long-term solution, but the U.S. in particular feared that Yeltsin's government would not survive a looming financial crisis without IMF help. U.S. President Bill Clinton's treasury secretary, Robert Rubin, also feared that a Russian collapse could create a panic on world money markets (and it indeed did help bring down one major US hedge fund Long-Term Capital Management). The IMF approved a $22.6 billion emergency loan on July 13.
Despite the bailout, Russia's monthly interest payments still well exceeded its monthly tax revenues. Realizing that this situation was unsustainable, investors continued to flee Russia despite the IMF bailout. Weeks later the financial crisis resumed and the value of the ruble resumed its fall, and the government fell into a self-perpetuating trap. To pay off the interest on the loans it had taken, it needed to raise still more cash, which it did through foreign borrowing. As lenders became increasingly certain that the government could not make good on its obligations, they demanded ever-higher interest rates, deepening the trap. Ultimately the bubble burst.
On August 17, Kiriyenko's government and the central bank were forced to suspend payment on Russia's foreign debt for 90 days, restructure the nation's entire debt, and devalue the ruble. The ruble went into free fall as Russians sought frantically to buy dollars. Western creditors lost heavily, and a large part of Russia's fledgling banking sector was destroyed, since many banks had substantial dollar borrowings. Foreign investment rushed out of the country, and financial crisis triggered an unprecedented flight of capital from Russia.
Political fallout
The financial collapse produced a political crisis, as Yeltsin, with his domestic support evaporating, had to contend with an emboldened opposition in the parliament. A week later, on August 23, Yeltsin fired Kiryenko and declared his intention of returning Chernomyrdin to office as the country slipped deeper into economic turmoil. Powerful business interests, fearing another round of reforms that might cause leading concerns to fail, welcomed Kiriyenko's fall, as did the Communists.
Yeltsin, who began to lose his hold as his health deteriorated, wanted Chernomyrdin back, but the legislature refused to give its approval. After the Duma rejected Chernomyrdin's candidacy twice, Yeltsin, his power clearly on the wane, backed down. Instead, he nominated Foreign Minister Yevgeny Primakov, who on September 11 was overwhelmingly approved by the Duma.
Primakov's appointment restored political stability because he was seen as a compromise candidate able to heal the rifts between Russia's quarreling interest groups. There was popular enthusiasm for Primakov as well. Primakov promised to make the payment of wage and pension arrears his government's first priority, and invited members of the leading parliamentary factions into his Cabinet.
Communists and trade unionists staged a nationwide strike on October 7, and called on President Yeltsin to resign. On October 9, Russia, which was also suffering from a bad harvest, appealed for international humanitarian aid, including food.
Recovery
Russia bounced back from the August 1998 financial crash with surprising speed. Much of the reason for the recovery is that world oil prices rapidly rose during 1999–2000 (just as falling energy prices on the world market had deepened Russia's financial troubles) so that Russia ran a large trade surplus in 1999 and 2000. Another reason is that domestic industries such as food processing have benefited from the devaluation, which caused a steep increase in the prices of imported goods. Also, since Russia's economy was operating to such a large extent on barter and other non-monetary instruments of exchange, the financial collapse had far less of an impact on many producers than it would had the economy been dependent on a banking system. Finally, the economy has been helped by an infusion of cash; as enterprises were able to pay off arrears in back wages and taxes, it, in turn, allowed consumer demand for the goods and services of Russian industry to rise. For the first time in many years, unemployment in 2000 fell as enterprises added workers.
Nevertheless, the political and social equilibrium of the country remains tenuous to this day, and power remains a highly personalized commodity. The economy remains vulnerable to downturn if, for instance, world oil prices fall at a dramatic pace.
Succession crisis
Yevgeny Primakov did not remain in his post long. Yeltsin grew suspicious that Primakov was gaining in strength and popularity and dismissed him in May 1999, after only eight months in office. Yeltsin then named Sergei Stepashin, who had formerly been head of the FSB (the successor agency to the KGB) and later been Interior Minister, to replace him. The Duma confirmed his appointment on the first ballot by a wide margin.
Stepashin's tenure was even shorter than Primakov's. In August 1999, Yeltsin once again abruptly dismissed the government and named Vladimir Putin as his candidate to head the new government. Like Stepashin, Putin had a background in the secret police, having made his career in the foreign intelligence service and later as head of the FSB. Yeltsin went so far as to declare that he saw Putin as his successor as president. The Duma narrowly voted to confirm Putin.
When appointed, Putin was a relatively unknown politician, but he quickly established himself both in public opinion and in Yeltsin's estimation as a trusted head of government, largely due to the Second Chechen War. Just days after Yeltsin named Putin as a candidate for prime minister, Chechen forces engaged the Russian army in Dagestan, a Russian autonomy near Chechnya. In the next month, several hundred people died in apartment building bombings in Moscow and other cities, bombings Russian authorities attributed to Chechen rebels. In response, the Russian army entered Chechnya in late September 1999, starting the Second Chechen War. The Russian public at the time, angry over the terrorist bombings, widely supported the war. The support translated into growing popularity for Putin, who had taken decisive action in Chechnya.
After the success of political forces close to Putin in the December 1999 parliamentary elections, Yeltsin evidently felt confident enough in Putin that he resigned from the presidency on December 31, six months before his term was due to expire. This made Putin acting president and gave Putin ample opportunity to position himself as the frontrunner for the Russian presidential election held on March 26, 2000, which he won. The Chechen War figured prominently in the campaign. In February 2000, Russian troops entered Grozny, the Chechen capital, and a week before the election, Putin flew to Chechnya on a fighter jet, claiming victory.
Putin era
In August 2000, the Russian submarine K-141 Kursk suffered an explosion, causing the submarine to sink in the shallow area of the Barents Sea. Russia organized a vigorous but hectic attempt to save the crew, and the entire futile effort was surrounded by unexplained secrecy. This, as well as the slow initial reaction to the event and especially to the offers of foreign aid in saving the crew, brought much criticism on the government and personally on President Putin.
On October 23, 2002, Chechen separatists took over a Moscow theater. Over 700 people inside were taken hostage in what has been called the Moscow theater hostage crisis. The separatists demanded the immediate withdrawal of Russian forces from Chechnya and threatened to blow up the building if authorities attempted to enter. Three days later, Russian commandos stormed the building after the hostages had been subdued with a sleeping gas, shooting the unconscious militants, and killing over 100 civilian hostages with the sleeping gas in the process.
In the aftermath of the theater siege, Putin began renewed efforts to eliminate the Chechen insurrection. (For additional details on the war in Chechnya under Putin, see Second Chechen War.) The government canceled scheduled troop withdrawals, surrounded Chechen refugee camps with soldiers, and increased the frequency of assaults on separatist positions.
Chechen militants responded in kind, stepping up guerrilla operations and rocket attacks on federal helicopters. Several high-profile attacks have taken place. In May 2004, Chechen separatists assassinated Akhmad Kadyrov, the pro-Russia Chechen leader who became the president of Chechnya 8 months earlier after an election conducted by Russian authorities. On August 24, 2004, two Russian aircraft were bombed. This was followed by the Beslan school hostage crisis in which Chechen separatists took 1,300 hostages. The initially high public support for the war in Chechnya has declined.
Putin has confronted several very influential oligarchs (Vladimir Gusinsky, Boris Berezovsky and Mikhail Khodorkovsky, in particular) who attained large stakes of state assets, allegedly through illegal schemes, during the privatization process. Gusinsky and Berezovsky have been forced to leave Russia and give up parts of their assets. Khodorkovsky was jailed in Russia and has lost his YUKOS company, formerly the largest oil producer in Russia. Putin's stand against oligarchs is generally popular with the Russian people, even though the jailing of Khodorkovsky was mainly seen as part of a takeover operation by government officials, according to another Levada-Center poll.
These confrontations have also led to Putin establishing control over Russian media outlets previously owned by the oligarchs. In 2001 and 2002, TV channels NTV (previously owned by Gusinsky), TV6 and TVS (owned by Berezovsky) were all taken over by media groups loyal to Putin. Similar takeovers have also occurred with print media.
Putin's popularity, which stems from his reputation as a strong leader, stands in contrast to the unpopularity of his predecessor, but it hinges on a continuation of economic recovery. Putin came into office at an ideal time: after the devaluation of the ruble in 1998, which boosted demand for domestic goods, and while world oil prices were rising. Indeed, during the seven years of his presidency, real GDP grew on average 6.7% a year, average income increased 11% annually in real terms, and a consistently positive balance of the federal budget enabled the government to cut 70% of the external debt (according to the Institute for Complex Strategic Studies). Thus, many credited him with the recovery, but his ability to withstand a sudden economic downturn has been untested. Putin won the Russian presidential election in March 2004 without any significant competition.
Some researchers assert that most Russians (as of 2007) have come to regret the collapse of the Soviet Union in 1991. On repeated occasions, even Vladimir Putin—Boris Yeltsin's handpicked successor — stated that the fall of Soviet rule had led to few gains and many problems for most Russian citizens. In a campaign speech in February 2004, for example, Putin called the dismantlement of the Soviet Union a "national tragedy on an enormous scale," from which "only the elites and nationalists of the republics gained." He added, "I think that ordinary citizens of the former Soviet Union and the post-Soviet space gained nothing from this. On the contrary, people have faced a huge number of problems."
Putin's international prestige suffered a major blow in the West during the disputed 2004 Ukrainian presidential election. Putin had twice visited Ukraine before the election to show his support for the pro-Russian Viktor Yanukovych against opposition leader Viktor Yushchenko, a pro-Western liberal economist. He congratulated Yanukovych, followed shortly afterwards by Belаrussian president Alexander Lukashenko, on his victory before election results were even made official and made statements opposing the rerun of the disputed second round of elections, won by Yanukovych, amid allegations of large-scale voting fraud. The second round was ultimately rerun; Yushchenko won the round and was eventually declared the winner on January 10, 2005. In the West, the reaction to Russia's handling of, or perhaps interference in, the Ukrainian election evoked echoes of the Cold War, but relations with the U.S. have remained stable.
On 14 March 2004, Putin was elected to the presidency for a second term, receiving 71% of the vote. The Beslan school hostage crisis took place in September 2004, in which hundreds died. Many in the Russian press and in the international media warned that the death of 130 hostages in the special forces' rescue operation during the 2002 Moscow theater hostage crisis would severely damage President Putin's popularity. However, shortly after the siege had ended, the Russian president enjoyed record public approval ratings – 83% of Russians declared themselves satisfied with Putin and his handling of the siege.
In 2005, the Russian government replaced the broad in-kind Soviet-era benefits, such as free transportation and subsidies for heating and other utilities for socially vulnerable groups by cash payments. The reform, known as monetization, has been unpopular and caused a wave of demonstrations in various Russian cities, with thousands of retirees protesting against the loss of their benefits. This was the first time such wave of protests took place during the Putin administration. The reform hurt the popularity of the Russian government, but Putin personally remained popular, with a 77% approval rating.
The near 10-year period prior to the rise of Putin after the dissolution of Soviet rule was a time of upheaval in Russia. In a 2005 Kremlin speech, Putin characterized the collapse of the Soviet Union as the "greatest geopolitical catastrophe of the Twentieth Century." Putin elaborated: "Moreover, the epidemic of disintegration infected Russia itself." The country's cradle-to-grave social safety net was gone and life expectancy declined in the period preceding Putin's rule. In 2005, the National Priority Projects were launched to improve Russia's health care, education, housing and agriculture.
The continued criminal prosecution of Russia's then richest man, President of Yukos oil and gas company Mikhail Khodorkovsky, for fraud and tax evasion was seen by the international press as a retaliation for Khodorkovsky's donations to both liberal and communist opponents of the Kremlin. The government said that Khodorkovsky was "corrupting" a large segment of the Duma to prevent changes to the tax code. Khodorkovsky was arrested, Yukos was bankrupted and the company's assets were auctioned at below-market value, with the largest share acquired by the state company Rosneft. The fate of Yukos was seen as a sign of a broader shift of Russia towards a system of state capitalism. This was underscored in July 2014 when shareholders of Yukos were awarded $50 billion in compensation by the Permanent Arbitration Court in The Hague.
On 7 October 2006, Anna Politkovskaya, a journalist who exposed corruption in the Russian army and its conduct in Chechnya, was shot in the lobby of her apartment building, on Putin's birthday. The death of Politkovskaya triggered international criticism, with accusations that Putin has failed to protect the country's new independent media. Putin himself said that her death caused the government more problems than her writings.
In 2007, "Dissenters' Marches" were organized by the opposition group The Other Russia, led by former chess champion Garry Kasparov and national-Bolshevist leader Eduard Limonov. Following prior warnings, demonstrations in several Russian cities were met by police action, which included interfering with the travel of the protesters and the arrests of as many as 150 people who attempted to break through police lines.
On 12 September 2007, Putin dissolved the government upon the request of Prime Minister Mikhail Fradkov. Fradkov commented that it was to give the President a "free hand" in the run-up to the parliamentary election. Viktor Zubkov was appointed the new prime minister.
In the December 2007 election, United Russia won 64.30% of the popular vote in their run for State Duma. This victory was seen by many as an indication of strong popular support of the then Russian leadership and its policies.
Putin was barred from a third term by the Constitution of Russia. First Deputy Prime Minister Dmitry Medvedev was elected his successor. In a power-switching operation on 8 May 2008, only a day after handing the presidency to Medvedev, Putin was appointed Prime Minister of Russia, maintaining his political dominance.
In 2008, Kosovo's declaration of independence saw a marked deterioration in Russia's relationship with the West. It also saw South Ossetia war against Georgia, that followed Georgia's attempt to take over the breakaway region of South Ossetia. Russian troops entered South Ossetia and forced Georgian troops back, establishing their control on this territory. In the fall of 2008, Russia unilaterally recognized the independence of South Ossetia and Abkhazia.
Putin has said that overcoming the consequences of the world economic crisis was one of the two main achievements of his second Premiership. The other was the stabilizing the size of Russia's population between 2008 and 2011 following a long period of demographic collapse that began in the 1990s.
At the United Russia Congress in Moscow on 24 September 2011, Medvedev officially proposed that Putin stand for the Presidency in March 2012, an offer Putin accepted. Given United Russia's near-total dominance of Russian politics, many observers believed that Putin was assured of a third term. The move was expected to see Medvedev stand on the United Russia ticket in the parliamentary elections in December, with a goal of becoming Prime Minister at the end of his presidential term.
After the parliamentary elections on 4 December 2011, tens of thousands of Russians engaged in protests against alleged electoral fraud, the largest protests in Putin's time. Protesters criticized Putin and United Russia and demanded annulment of the election results. Those protests sparked the fear of a colour revolution in society. Putin allegedly organized a number of paramilitary groups loyal to himself and to the United Russia party in the period between 2005 and 2012.
On 4 March 2012, Putin won the 2012 Russian presidential elections in the first round, with 63.6% of the vote, despite widespread accusations of vote-rigging. Opposition groups accused Putin and the United Russia party of fraud. While efforts to make the elections transparent were publicized, including the usage of webcams in polling stations, the vote was criticized by the Russian opposition and by international observers from the Organization for Security and Co-operation in Europe for procedural irregularities.
Anti-Putin protests took place during and directly after the presidential campaign. The most notorious protest was the Pussy Riot performance on 21 February, and subsequent trial. An estimated 8,000–20,000 protesters gathered in Moscow on 6 May, when eighty people were injured in confrontations with police, and 450 were arrested, with another 120 arrests taking place the following day.
In 2012 and 2013, Putin and the United Russia party backed stricter legislation against the LGBT community, in Saint Petersburg, Archangelsk and Novosibirsk; a law called the Russian gay propaganda law, that is against "homosexual propaganda" (which prohibits such symbols as the rainbow flag as well as published works containing homosexual content) was adopted by the State Duma in June 2013. Responding to international concerns about Russia's legislation, Putin asked critics to note that the law was a "ban on the propaganda of pedophilia and homosexuality" and he stated that homosexual visitors to the 2014 Winter Olympics should "leave the children in peace" but denied there was any "professional, career or social discrimination" against homosexuals in Russia.
Ukrainian war
On 22 February 2014, the Yanukovych government of Ukraine collapsed as a result of the 2014 Ukrainian revolution, which the Russian government called a foreign backed proxy movement. On the same day, according to Russian president Vladimir Putin, he called an all-night meeting of his military leaders, at the end of which he ordered them to “begin the work to bring Crimea back into Russia.” By February 27, unmarked Russian troops in Ukraine were establishing a blockade of the borders and Ukrainian military bases in the Autonomous Republic of Crimea, and took armed control of its regional parliament.
A new Ukrainian government was formed and scheduled new elections for May 2014. On 1 March, from exile, Viktor Yanukovych requested that Russia use military forces "to establish legitimacy, peace, law and order, stability and defending the people of Ukraine". On the same day, Vladimir Putin requested and received authorization from the Russian Parliament to deploy Russian troops to Ukraine in response to the crisis and gained complete control over Crimean Peninsula within a day.
On 6 March 2014, the Crimean Parliament voted to "enter into the Russian Federation with the rights of a subject of the Russian Federation" and later held a referendum asking the people of these regions whether they wanted to join Russia as a federal subject, or if they wanted to restore the 1992 Crimean constitution and Crimea's status as a part of Ukraine. Though passed with an overwhelming majority, the results are contested by some and approved by others.
Crimea and Sevastopol formally declared independence as the Republic of Crimea and requested that they be admitted as constituents of the Russian Federation. On 18 March 2014, Russia and Crimea signed a treaty of accession of the Republic of Crimea and Sevastopol in the Russian Federation, though the United Nations General Assembly voted in favor of a non-binding statement to oppose Russia's annexation of the peninsula.
In 2022 the crisis escalated to an all-out invasion of Ukraine.
Relations with the West
In the early period after Russia became independent, Russian foreign policy repudiated Marxism–Leninism as a putative guide to action, emphasizing cooperation with the West in solving regional and global problems, and soliciting economic and humanitarian aid from the West in support of internal economic reforms.
However, although Russia's leaders now described the West as its natural ally, they grappled with defining new relations with the East European states, the new states formed upon the disintegration of Yugoslavia, and Eastern Europe. Russia opposed the expansion of NATO into the former Soviet bloc nations of the Czech Republic, Poland, and Hungary in 1997 and, particularly, the second NATO expansion into the Baltic states in 2004. In 1999, Russia opposed the NATO bombing of Yugoslavia for more than two months (see Kosovo War), but later joined NATO peace-keeping forces in the Balkans in June 1999.
Relations with the West have also been stained by Russia's relationship with Belarus. Belorussian President Alexander Lukashenko, an authoritarian leader, has shown much interest in aligning his country with Russia, and no interest in deepening ties with the ever-expanding NATO or implementing Western-backed neoliberal economic reforms. A union agreement between Russia and Belarus was formed on April 2, 1996. The agreement was tightened, becoming the Union of Russia and Belarus on April 3, 1997. Further strengthening of the union occurred on December 25, 1998, and in 1999.
Under Putin, Russia has sought to strengthen ties with the People's Republic of China by signing the Treaty of Good-Neighborliness and Friendly Cooperation as well building the Trans-Siberian oil pipeline geared toward growing Chinese energy needs. He also made a number of appearances in the media with President of the United States George W. Bush in which the two described each other as "friends".
See also
Economy of Russia
Politics of Russia
Timeline of largest projects in the Russian economy
Notes
References
Further reading
Rosefielde, Steven. Putin's Russia: Economy, Defence and Foreign Policy (2020) excerpt
External links
Charles H. Fairbanks, Jr., "What Went Wrong in Russia?: The Feudalization of the State," Journal of Democracy (April 1999), p. 47–53. (an analysis of the state of democratic transition in Russia by Charles H. Fairbanks, Jr., a leading specialist in Russian politics published in the Journal of Democracy)
Post-Soviet Russia and Its Neighbors from the Dean Peter Krogh Foreign Affairs Digital Archives
Paul J. Saunders, "Why "Globalization" Didn't Rescue Russia" (Paul J. Saunders is a specialist in U.S.-Russian relations and executive director of the Center for the National Interest)
Up for Debate: Shock Therapy: Bolivia, Poland, Russia. Same Policies-Different Results From the PBS series "Commanding Heights"
Up for Debate: Privatization: Who Wins? Russia's Reform Compromise (interviews with Yegor Gaidar, Grigory Yavlinsky, Anatoly Chubais, Joseph Stiglitz, and Jeffrey Sachs from the PBS series "Commanding Heights")
Nick Paton Walsh, The Guardian, July 6, 2005, "Meet the chief exec of Kremlin inc ..."
Failed Expectations, Or What Is Behind the Marshall Plan for Post-Socialist Reconstruction'', by Tanya Narozhna
"The Uncertain Return of Russian Power", by Mike Haynes, analyses Putin's Russia and looks at claims that the reassertion of Russian power is leading to a new Cold War.
Mike Edwards: "Russia — Playing by new rules" National Geographic Magazine March 1993
Mark Hollingsworth & Stewart Lansley, Londongrad: From Russia With Cash, 2009, 4th Estate
1991 establishments in Russia
1991
1991
1991
Russia |
null | null | Town | eng_Latn | A town is a human settlement. Towns are generally larger than villages and smaller than cities, though the criteria to distinguish between them vary considerably in different parts of the world.
Origin and use
The word "town" shares an origin with the German word , the Dutch word , and the Old Norse . The original Proto-Germanic word, *tūnan, is thought to be an early borrowing from Proto-Celtic *dūnom (cf. Old Irish , Welsh ).
The original sense of the word in both Germanic and Celtic was that of a fortress or an enclosure. Cognates of town in many modern Germanic languages designate a fence or a hedge. In English and Dutch, the meaning of the word took on the sense of the space which these fences enclosed, and through which a track must run. In England, a town was a small community that could not afford or was not allowed to build walls or other larger fortifications, and built a palisade or stockade instead. In the Netherlands, this space was a garden, more specifically those of the wealthy, which had a high fence or a wall around them (like the garden of the palace of Het Loo in Apeldoorn, which was the model for the privy garden of William III and Mary II at Hampton Court). In Old Norse means a (grassy) place between farmhouses, and the word is still used with a similar meaning in modern Norwegian.
Old English became a common place-name suffix in England and southeastern Scotland during the Anglo-Saxon settlement period. In Old English and Early and Middle Scots, the words ton, toun, etc. could refer to diverse kinds of settlements from agricultural estates and holdings, partly picking up the Norse sense (as in the Scots word ) at one end of the scale, to fortified municipalities. Other common Anglo-Saxon suffixes included ham 'home', stede 'stead', and burh 'bury, borough, burgh'.
In some cases, town is an alternative name for "city" or "village" (especially a larger village). Sometimes, the word town is short for township. In general, today towns can be differentiated from townships, villages, or hamlets on the basis of their economic character, in that most of a town's population will tend to derive their living from manufacturing industry, commerce, and public services rather than primary sector industries such as agriculture or related activities.
A place's population size is not a reliable determinant of urban character. In many areas of the world, e.g. in India at least until recent times, a large village might contain several times as many people as a small town. In the United Kingdom, there are historical cities that are far smaller than the larger towns.
The modern phenomenon of extensive suburban growth, satellite urban development, and migration of city dwellers to villages has further complicated the definition of towns, creating communities urban in their economic and cultural characteristics but lacking other characteristics of urban localities.
Some forms of non-rural settlement, such as temporary mining locations, may be clearly non-rural, but have at best a questionable claim to be called a town.
Towns often exist as distinct governmental units, with legally defined borders and some or all of the appurtenances of local government (e.g. a police force). In the United States these are referred to as "incorporated towns". In other cases the town lacks its own governance and is said to be "unincorporated". Note that the existence of an unincorporated town may be legally set out by other means, e.g. zoning districts. In the case of some planned communities, the town exists legally in the form of covenants on the properties within the town. The United States Census identifies many census-designated places (CDPs) by the names of unincorporated towns which lie within them; however, those CDPs typically include rural and suburban areas and even surrounding villages and other towns.
The distinction between a town and a city similarly depends on the approach: a city may strictly be an administrative entity which has been granted that designation by law, but in informal usage, the term is also used to denote an urban locality of a particular size or importance: whereas a medieval city may have possessed as few as 10,000 inhabitants, today some consider an urban place of fewer than 100,000 as a town, even though there are many officially designated cities that are much smaller than that.
In toponymic terminology, names of individual towns and cities are called astyonyms or astionyms (from Ancient Greek 'town, city', and 'name').
History
Through different periods of recorded history, many towns have grown into sizeable settlements, with the development of properties, centres of culture, and specialized economies.
Neolithic
Çatalhöyük, currently an archaeological site, was considered to be the oldest inhabited town that existed from around 7500 BC. Inscribed as a world heritage site, it remains a depopulated town with a complex of ruins.
Roman era
In Roman times, a villa was a rural settlement formed by a main residential building and another series of secondary buildings. It constituted the center from which an agricultural holding was administered. Subsequently, it lost its agricultural functions and reduced its activity to residential. With the consolidation of large estates during the Roman Empire, the town became the center of large farms.
A distinction was created between rustic and urban settlements:
Rustic villas, from where the exploitation of resources was directed, slave workers resided, livestock were kept and production was stored.
Urban villas, in which the lord resided and which increasingly adopted the architectural and beautification forms typical of urban mansions. When from the first century the great territorial property was divided between the area directly exploited by the lord and that ceded to tenant settlers, urban villas became the centers of the administrative power of the lords, appearing the forms of vassalage typical of feudalism of the fourth century.
Terminology for statistics
193 countries have been involved in a common effort to agree on a common statistical definition of the three categories: cities, towns and rural areas.
Age of towns scheme
Australian geographer Thomas Griffith Taylor proposed a classification of towns based on their age and pattern of land use. He identified five types of town:
Infantile towns, with no clear zoning
Juvenile towns, which have developed an area of shops
Adolescent towns, where factories have started to appear
Early mature towns, with a separate area of high-class housing
Mature towns, with defined industrial, commercial and various types of residential area
By country
Afghanistan
In Afghanistan, towns and cities are known as shār (Dari: شهر, Pashto: ښار). As the country is an historically rural society with few larger settlements, with major cities never holding more than a few hundred thousand inhabitants before the 2000s, the lingual tradition of the country does not discriminate between towns and cities.
Albania and Kosovo
In Albania and Kosovo means 'town', which is very similar to the word for city (), although there is no official use of the term for any settlement.
In Albanian means 'small city' or 'new city', while in ancient times it referred to a small residential center within the walls of a castle.
Australia
In Australia, most rural and regional centres of population can be called towns; many small towns have populations of less than 200. The smallest may be described as townships.
In addition, some local government entities are officially styled as towns in Queensland, South Australia, Western Australia and the Northern Territory, and formerly also (till the 1990s) in Victoria.
Austria
The Austrian legal system does not distinguish between villages, towns, and cities. The country is partitioned into 2098 municipalities () of fundamentally equal rank. Larger municipalities are designated as market towns () or cities (), but these distinctions are purely symbolic and do not confer additional legal responsibilities. There is a number of smaller communities that are labelled cities because they used to be regional population centers in the distant past. The city of Rattenberg for example has about 400 inhabitants. The city of Hardegg has about 1200 inhabitants, although the historic city core − Hardegg proper without what used to be the surrounding hamlets − is home to just 80 souls.
There are no unincorporated areas.
Of the 201 cities in Austria, 15 are statutory cities (). A statutory city is a city that is vested, in addition to its purview as a municipality, with the duties of a district administrative authority. The status does not come with any additional autonomy: district administrative authorities are essentially just service centers that citizens use to interact with the national government, for example to apply for driver licenses or passports. The national government generally uses the provinces to run these points of contact on its behalf; in the case of statutory cities, the municipality gets to step up.
Bulgaria
Bulgarians do not, in general, differentiate between 'city' and 'town'. However, in everyday language and media the terms "large towns" and "small towns" are in use. "Large towns" usually refers to Sofia, Plovdiv, Varna and Burgas, which have population over 200,000. Ruse and Stara Zagora are often included as well due to presence of relatively developed infrastructure and population over 100,000 threshold. It is difficult to call the remaining provincial capitals "large towns" as, in general, they are less developed and have shrinking population, some with as few as 30,000 inhabitants.
In Bulgaria the Council of Ministers defines what constitutes a settlement, while the President of Bulgaria grants each settlement its title. In 2005 the requirement that villages that wish to classify themselves as town must have a social and technical infrastructure, as well as a population of no fewer than 3500 people. For resort settlements the requirements are lower with the population needing to be no fewer than 1000 people but infrastructure requirements remain.
Canada
The legal definition of a town in Canada varies by province or territory, as each has jurisdiction over defining and legislating towns, cities and other types of municipal organization within its own boundaries.
The province of Quebec is unique in that it makes no distinction under law between towns and cities. There is no intermediate level in French between and (municipality is an administrative term usually applied to a legal, not geographical entity), so both are combined under the single legal status of ville. While an informal preference may exist among English speakers as to whether any individual is commonly referred to as a city or as a town, no distinction and no objective legal criteria exist to make such a distinction under law.
Chile
In Chile, towns (Spanish: ) are defined by the National Statistics Institute (INE) as an urban entity with a population from 2001 to 5000 or an area with a population from 1001 to 2000 and an established economic activity.
Czech Republic
In the Czech Republic, a municipality can obtain the title of a city (), town () or market town (). The title is granted by law.
Statutory cities (in English usually called just "cities"), which are defined by law no. 128/2000 Coll., can define their own self-governing municipal districts. There are 26 such cities, in addition to Prague, which is a de facto statutory city. All the Czech municipalities with more than 40,000 inhabitants are cities.
Town and market town are above all ceremonious honorary degrees, referring to population, history and regional significance of a municipality. As the statistics of Czech municipalities shows, towns usually have between 1,000 and 35,000 inhabitants, with median around 4,000 and average around 6,500. Nowadays a municipality must have at least 3,000 inhabitants to have the right to request the town title. Market towns usually have between 500 and 4,000 inhabitants, with median and average both around 1,000.
Denmark
In Denmark, in many contexts no distinction is made between "city", "town" and "village"; all three translate as . In more specific use, for small villages and hamlets the word (meaning 'country town') is used, while the Danish equivalent of English city is (meaning 'large town'). For formal purposes, urban areas having at least 200 inhabitants are considered .
Historically some towns held various privileges, the most important of which was the right to hold market. They were administered separately from the rural areas in both fiscal, military and legal matters. Such towns are known as (roughly the same meaning as borough albeit deriving from a different etymology) and they retain the exclusive right to the title even after the last vestiges of their privileges vanished through the reform of the local administration carried through in 1970.
Estonia
In Estonia, there is no distinction between a town and a city as the word is used for both bigger and smaller settlements, which are bigger than villages and boroughs. There are 30 municipal towns () in Estonia and a further 17 towns, which have merged with a municipal parish ().
Finland
In Finland, there is no distinction between a town and a city as the word is used for both bigger and smaller settlements, which are bigger than villages and boroughs; although when talking about the word town, the word is used ( means 'little' or 'small'). There are almost one hundred municipal towns in Finland.
France
From an administrative standpoint, the smallest level of local authorities are all called communes. They can have anywhere from a handful to millions of inhabitants, and France has 36,000 of them. The French term for town is bourg but French laws generally do not distinguish between towns and cities which are all commonly called . However, some laws do treat these authorities differently based on the population and different rules apply to the three big cities Paris, Lyon and Marseille. For historical reasons, six communes in the Meuse département exist as independent administrative entities despite having no inhabitants at all.
For statistical purposes, the national statistical institute (INSEE) operates a distinction between urban areas with fewer than 2,000 inhabitants and bigger communes, the latter being called . Smaller settlements are usually called .
Germany
Germans do not, in general, differentiate between 'city' and 'town'. The German word for both is , as it is the case in many other languages that do not differentiate between these concepts. The word for a 'village', as a smaller settlement, is . However, the International Statistics Conference of 1887 defined different sizes of , based on their population size, as follows: ('country town'; under 5,000), ('small town'; 5,000 to 20,000), ('middle town'; between 20,000 and 100,000) and ("large town"; 100,000 to 1,000,000). The term may be translated as 'city'. In addition, Germans may speak of a , a city with anywhere between one and five million inhabitants (such as Cologne, Munich, Hamburg and Berlin). Also, a city with more than five million inhabitants is often referred to as a (commonly translated as Megacity).
Historically, many settlements became a by being awarded a Stadtrecht in medieval times. In modern German language use, the historical importance, the existence of central functions (education, retail etc.) and the population density of an urban place might also be taken as characteristics of a . The modern local government organisation is subject to the laws of each state and refers to a (municipality), regardless of its historic title. While most form part of a (district) on a higher tier of local government, larger towns and cities may have the status of a , combining both the powers of a municipality and a district.
Designations in different states are as diverse as e.g. in Australian States and Territories, and differ from state to state. In some German states, the words ('market'), (both used in southern Germany) or ('spot'; northern Germany e.g. in Lower Saxony) designate a town-like residential community between and with special importance to its outer conurbation area. Historically those had (market right) but not full town privileges; see Market town. The legal denomination of a specific settlement may differ from its common designation (e.g. – a legal term in Lower Saxony for a group of villages [, pl. ] with common local government created by combining municipalities [, pl. ]).
Greece and Cyprus
In ordinary speech, Greeks use the word ('village') to refer to smaller settlements and the word or ('city') to refer to larger ones. Careful speakers may also use the word to refer to towns with a population of 2,000–9,999.
In Greek administrative law there used to be a distinction between , i.e. municipalities with more than 10,000 inhabitants or considered important for some other geographical (county seats), historical or ecclesiastical (bishops' seats) reason, and κοινότητες, referring to smaller self-governing units, mostly villages. A sweeping reform, carried out in two stages early in the 21st century, merged most with the nearest , dividing the whole country into 325 big self-governing . The former municipalities survive as administrative subdivisions (, ).
Cyprus, including the Turkish-occupied areas, is also divided into 39 (in principle, with at least 5,000 inhabitants, though there are exceptions) and 576 .
Hong Kong
Hong Kong started developing new towns in the 1950s, to accommodate exponential population increase. The first new towns included Tsuen Wan and Kwun Tong. In the late 1960s and the 1970s, another stage of new town developments was launched. Nine new towns have been developed so far. Land use is carefully planned and development provides plenty of room for public housing projects. Rail transport is usually available at a later stage. The first towns are Sha Tin, Tsuen Wan, Tuen Mun and Tseung Kwan O. Tuen Mun was intended to be self-reliant, but was not successful and turned into a bedroom community like the other new towns. More recent developments are Tin Shui Wai and North Lantau (Tung Chung-Tai Ho).
Hungary
In Hungary there is no official distinction between a city and a town (the word for both in Hungarian is ). Nevertheless, the expressions formed by adding the adjectives ('small') and ('large') to the beginning of the root word (e.g. ) have been normalized to differentiate between cities and towns (towns being smaller, therefore bearing the name .) In Hungary, a village can gain the status of ('town'), if it meets a set of diverse conditions for quality of life and development of certain public services and utilities (e.g. having a local secondary school or installing full-area sewage collection pipe network). Every year the Minister of Internal Affairs selects candidates from a committee-screened list of applicants, whom the President of Republic usually affirms by issuing a bill of town's rank to them. Since being a town carries extra fiscal support from the government, many relatively small villages try to win the status of ('town rank') nowadays.
Before the fall of communism in 1990, Hungarian villages with fewer than 10,000 residents were not allowed to become towns. Recently some settlements as small as 2,500 souls have received the rank of town (e.g. Visegrád, Zalakaros or Gönc) and meeting the conditions of development is often disregarded to quickly elevate larger villages into towns. As of middle 2013, there are 346 towns in Hungary, encompassing some 69% of the entire population.
Towns of more than 50,000 people are able to gain the status of (town with the rights of a county), which allows them to maintain a higher degree of services. (There are a few exceptions, when towns of fewer than 50,000 people gained the status: Érd, Hódmezővásárhely, Salgótarján and Szekszárd) As of middle 2013, there are only 23 such towns in Hungary.
Iceland
Ireland
The Local Government act 2001 provides that from January 1, 2002 (section 10 subsection (3)
Within the county in which they are situated and of which they form part, there continue to be such other local government areas as are set out in Schedule 6 which – (a) in the case of the areas set out in Chapter 1 of Part 1 of that Schedule, shall be known as boroughs, and – (b) in the case of the areas set out in Chapter 2 of Part 1 and Part 2 of that Schedule, shall be known as towns, and in this Act a reference to a town shall include a reference to a borough.
These provisions affect the replacement of the boroughs, towns and urban districts which existed before then. Similar reforms in the nomenclature of local authorities ( but not their functions) are effected by section 11 part 17 of the act includes provision (section 185(2))
Qualified electors of a town having a population of at least 7,500 as ascertained at the last preceding census or such other figure as the Minister may from time to time prescribe by regulations, and not having a town council, may make a proposal in accordance with paragraph (b) for the establishment of such a council
and contains provisions enabling the establishment of new town councils and provisions enabling the dissolution of existing or new town councils in certain circumstances
The reference to town having a population of at least 7,500 as ascertained at the last preceding census hands much of the power relating to defining what is in fact a town over to the Central Statistics Office and their criteria are published as part of each census.
Planning and Development Act 2000
Another reference to the Census and its role in determining what is or is not a town for some administrative purpose is in the Planning and Development act 2000 (part II chapter I which provides for Local area plans)
A local area plan shall be made in respect of an area which —(i) is designated as a town in the most recent census of population, other than a town designated as a suburb or environs in that census, (ii) has a population in excess of 2,000, and (iii) is situated within the functional area of a planning authority which is a county council.
Central Statistics Office criteria
These are set out in full at 2006 Census Appendices.
In short they speak of "towns with legally defined boundaries" ( i.e. those established by the Local Government Act 2001) and the remaining 664 as "census towns", defined by themselves since 1971 as a cluster of 50 or more occupied dwellings in which within a distance of 800 meters there is a nucleus of 30 occupied houses on both sides of the road or twenty occupied houses on one side of the road there is also a 200 meter criterion for determining whether a house is part of a census town.
India
The 2011 Census of India defines towns of two types: statutory town and census town. Statutory town is defined as all places with a municipality, corporation, cantonment board or notified town area committee. Census towns are defined as places that satisfy the following criteria:
Minimum population of 5,000
At least 75% of male working population engaged in non-agricultural pursuits
Density of population at least 400/km2. (1,000 per sq. mile).
All the statutory towns, census towns and out growths are considered as urban settlements, as opposed to rural areas.
Towns in India usually have basic infrastructure like shops, electricity, bituminised roads, post offices, banks, telephone facilities, high schools and sometimes a few government offices. The human population living in these towns may be a few thousand. There are some towns which can be labelled as Main road town.
In state of Karnataka, towns are known as or in the Kannada language. Sometimes the terms ('city') or , which generally means 'place', are used for towns. The administrative council which governs these towns is known as or in Kannada depending on the number of people living within the town's boundaries.
Iran
In contemporary Persian texts, no distinction is made between city and town; both translate as (). In older Persian texts (until the first half of the 20th century), the Arabic word () was used for a town. However, in the past 50 years, this word has become obsolete.
There is a word in Persian which is used for special sort of satellite townships and city neighborhoods. It is (), (lit.: 'small city').
Another smaller type of town or neighborhood in a big city is called (). and each have different legal definitions.
Large cities such as Tehran, Mashhad, Isfahan, Tabriz, etc. which have millions inhabitants are referred to as (), metropole.
The pace in which different large villages have gained city status in Iran shows a dramatic increase in the last two decades.
Bigger cities and towns usually are centers of a township (in Persian: (). itself is a subdivision of (), 'province'.
Iraq
The word () is used to describe villages, the word () to describe towns, and the word () to describe cities.
Isle of Man
There are four settlements which are historically and officially designated as towns (Douglas, Ramsey, Peel, Castletown); however
Peel is also sometimes referred to as a city by virtue of its cathedral.
Onchan and Port Erin are both larger in population than the smallest "town", having expanded in modern times, but are designated as villages.
Israel
Modern Hebrew does provide a word for the concept of a town: (), derived from (), the biblical word for 'city'. However, the term is normally used only to describe towns in foreign countries, i.e. urban areas of limited population, particularly when the speaker is attempting to evoke nostalgic or romantic attitudes. The term is also used to describe a Shtetl, a pre-Holocaust Eastern European Jewish town.
Within Israel, established urban areas are always referred to as cities (with one notable exception explained below) regardless of their actual size. Israeli law does not define any nomenclature for distinction between urban areas based on size or any other factor – meaning that all urban settlements in Israel are legally referred to as "cities".
The exception to the above is the term (, lit. 'Development Town') which is applied to certain cities in Israel based on the reasons for their establishment. These cities, created during the earlier decades of Israeli independence (1950s and 1960s, generally), were designed primarily to serve as commercial and transportation hubs, connecting smaller agricultural settlements in the northern and southern regions of the country (the "Periphery") to the major urban areas of the coastal and central regions. Some of these development towns have since grown to a comparatively large size, and yet are still referred to as development towns, particularly when the speaker wishes to emphasize their (often low) socio-economic status. Nonetheless, they are rarely (if ever) referred to simply as towns; when referring to one directly, it will be called either a development town or a city, depending on context.
Italy
Although Italian provides different words for city (), town () and village (, old-fashioned, or frazione, most common), no legal definitions exist as to how settlements must be classified. Administratively, both towns and cities are ruled as comuni/comunes, while villages might be subdivisions of the former.
Generally, in everyday speech, a town is larger or more populated than a village and smaller than a city. Various cities and towns together may form a metropolitan area (). A city, can also be a culturally, economically or politically prominent community with respect to surrounding towns. Moreover, a city can be such by Presidential decree. A town, in contrast, can be an inhabited place which would elsewhere be styled a city, but has not received any official recognition.
Remarkable exceptions do exist: for instance, Bassano del Grappa, was given the status of in 1760 by Francesco Loredan's dogal decree and has since then carried this title. Also, the Italian word for 'town' ( with lowercase P) must not be confused with the Italian word for 'country/nation' ( usually with uppercase P).
Japan
In Japan city status (shi) was traditionally reserved for only a few particularly large settlements. Over time however the necessary conditions to be a city have been watered down and today the only loose rules that apply are having a population over 50,000 and over 60% of the population in a "city centre". In recent times many small villages and towns have merged in order to form a city despite seeming geographically to be just a collection of villages.
The distinction between towns (machi/chō) and villages (mura/son) is largely unwritten and purely one of population size when the settlement was founded with villages having under 10,000 and towns 10,000–50,000.
Korea
In both of South Korea and North Korea, towns are called eup ().
Latvia
In Latvia, towns and cities are indiscriminately called in singular form. The name is a contraction of two Latvian words: ('castle') and ('fence'), making it very obvious what is meant by the word – what is situated between the castle and the castle fence. However, a city can be called in reference to its size. A village is called or in Latvian.
Lithuania
In Lithuanian, a city is called miestas and a town is called miestelis (literally 'small '). Metropolises are called didmiestis (literally 'big ').
Malaysia
In Malaysia, a town is the area administered by a municipal council ().
Netherlands
Before 1848 there was a legal distinction between and non- parts of the country, but the word no longer has any legal significance. About 220 places were granted ('city rights') and are still so called for historical and traditional reasons, though the word is also used for large urban areas that never obtained such rights. Because of this, in the Netherlands, no distinction is made between city and town; both translate as . A hamlet () usually has fewer than 1,000 inhabitants, a village () ranges from 1,000 up to 25,000 inhabitants, and a place above 25,000 can call itself either village or city, mostly depending on historic reasons or size of the place. As an example, The Hague never gained city rights, but because of its size – more than half a million inhabitants – it is regarded as a city. Staverden, with only 40 inhabitants, would be a hamlet, but because of its city rights it may call itself a city.
For statistical purposes, the Netherlands has three sorts of cities:
(small city): 50,000–99,999 inhabitants
(medium-sized city): 100,000–249,999 inhabitants
(large city): 250,000 or more
Only Amsterdam, Rotterdam, The Hague and Utrecht are regarded as a .
New Zealand
In New Zealand, a town is a built-up area that is not large enough to be considered a city. Historically, this definition corresponded to a population of between approximately 1,000 and 20,000. Towns have no independent legal existence, being administered simply as built-up parts of districts, or, in some cases, of cities.
New Zealand's towns vary greatly in size and importance, ranging from small rural service centres to significant regional centres such as Blenheim and Taupo. Typically, once a town reaches a population of somewhere between 20,000 and 30,000 people, it will begin to be informally regarded as a city. One who regards a settlement as too small to be a town will typically call it a "township" or "village."
Norway
In Norway, city and town both translate to by, even if a city may be referred to as ('large town'). They are all part of and administered as a kommune ('municipality').
Norway has had inland the northernmost city in the world: Hammerfest. Now the record is held by New Ålesund on the Norwegian island Svalbard.
Philippines
In the Philippines, the local official equivalent of the town is the municipality (Filipino: ). Every municipality, or town, in the country has a mayor () and a vice mayor () as well as local town officials (Sangguniang Bayan). Philippine towns, otherwise called municipalities, are composed of a number of villages and communities called barangays with one (or a few cluster of) (s) serving as the town center or poblacion.
Unique in Philippine towns is that they have fixed budget, population and land requirements to become as such, i.e. from a , or a cluster of such, to a town, or to become cities, i.e. from town to a city. Respectively, examples of these are the town of B.E. Dujali in Davao del Norte province, which was formed in 1998 from a cluster of five , and the city of El Salvador, which was converted from a town to a city in 2007. Each town in the Philippines was classified by its annual income and budget.
A sharp, hierarchical distinction exists between Philippine cities ( or ) and towns, as towns in the country are juridically separate from cities, which are typically larger and more populous (some smaller and less populated) and which political and economic status are above those of towns. This was further supported and indicated by the income classification system implemented by the National Department of Finance, to which both cities and towns fell into their respective categories that indicate they are such as stated under Philippine law. However, both towns and cities equally share the status as local government units (LGUs) grouped under and belong to provinces and regions; both each are composed of and are governed by a mayor and a vice mayor supplemented by their respective LGU legislative councils. However despite this some towns in the Philippines are significantly larger than some cities in the Philippines such as Rodriguez, Rizal, Santa Maria, Bulacan and Minglanilla, Cebu are actually bigger than some regional centers.
Poland
In the Polish language there is no linguistic distinction between a city and a town. The word for both is , as a form of settlement distinct from following: village (), hamlet (), settlement (), or colony (). Town status is conferred by administrative decree; new towns are announced by the government in a separate law effective from the first day of the year. Some settlements tend to remain villages even though they have a larger population than many smaller towns. Towns may be called , a diminutive colloquially used for localities with a few thousand residents. Such localities have usually a mayor () as the head of the town council.
Cities are the biggest localities and generally must be bigger than 100,000 residents. They are run by the president () as the head of the city council. There are a few (mainly historic or political) exemptions which have allowed towns with fewer than 100,000 people to obtain the title of president for their mayors and to become recognized as cities that way, including Bolesławiec, Gniezno, and Zamość.
Portugal
Like other Iberian cultures, in Portugal there is a traditional distinction between towns (vilas) and cities (cidades). Similarly, although these areas are not defined under the constitution, and have no political function (with associated organs), they are defined by law, and a town must have:
at least 3,000 voters
at least half of these services: health unit, pharmacy, cultural centre, public transportation network, post office, commercial food and drinking establishments, primary school and/or bank office
In this context, the town or city is subordinate to the local authority (civil parish or municipality, in comparison to the North American context, where they have political functions. In special cases, some villages may be granted the status of town if they possess historical, cultural or architectonic importance.
The Portuguese urban settlements heraldry reflects the difference between towns and cities, with the coat of arms of a town bearing a crown with four4 towers, while the coat of arms of a city bears a crown withfive5 towers. This difference between towns and cities is still in use in other Portugues- speaking countries, but in Brazil is no longer in use.
Romania
In Romania there is no official distinction between a city and a town (the word for both in Romanian is ). Cities and towns in Romania can have the status either of municipiu, conferred to large urban areas, or only to smaller urban localities. Some settlements remain villages (comune) even though they have a larger population than other smaller towns.
Russia
Unlike English, the Russian language does not distinguish the terms city and town—both are translated as (). Occasionally the term is applied to urban-type settlements as well, even though the status of those is not the same as that of a city/town proper.
In Russia, the criteria an inhabited locality needs to meet in order to be granted city/town () status vary in different federal subjects. In general, to qualify for this status, an inhabited locality should have more than 12,000 inhabitants and the occupation of no less than 85% of inhabitants must be other than agriculture. However, inhabited localities which were previously granted the city/town status but no longer meet the criteria can still retain the status for historical reasons.
Singapore
In Singapore, towns are large scale satellite housing developments which are designed to be self-contained. It includes public housing units, a town centre and other amenities. Helmed by a hierarchy of commercial developments, ranging from a town centre to precinct-level outlets, there is no need to venture out of town to meet the most common needs of residences. Employment can be found in industrial estates located within several towns. Educational, health care, and recreational needs are also taken care of with the provision of schools, hospitals, parks, sports complexes, and so on. The most populous town in the country is Bedok.
South Africa
In South Africa the Afrikaans term is used interchangeably with the English equivalent town. A town is a settlement that has a size that is smaller than that of a city.
Spain
In Spain, the equivalent of town would be , a population unit between a village () and a city (), and is not defined by the number of inhabitants, but by some historical rights and privileges dating from the Middle Ages, such as the right to hold a market or fair. For instance, while Madrid is technically a , Barcelona, with a smaller population, is known as a city.
Sweden
The Swedish language does not differentiate between towns and cities in the English sense of the words; both words are commonly translated as , a term which has no legal significance today. The term is used for an urban area or a locality, which however is a statistical rather than an administrative concept and encompasses densely settled villages with only 200 inhabitants as well as the major cities. The word köping corresponds to an English market town (chipping) or German Markt but is mainly of historical significance, as the term is not used today and only survives in some toponyms. Some towns with names ending in are cities with over 100,000 inhabitants today, e.g. Linköping.
Before 1971, 132 larger municipalities in Sweden enjoyed special royal charters as stad instead of kommun (which is similar to a US county). However, since 1971 all municipalities are officially defined as , thus making no legal difference between, for instance, Stockholm and a small countryside municipality. Every urban area that was a before 1971 is still often referred to as a in daily speech. Since the 1980s, 14 of these municipalities have branded themselves as again, although this has no legal or administrative significance, as they still have to refer to themselves as in all legal documentation.
For statistical purposes, Statistics Sweden officially defines a as an urban area of at least 10,000 inhabitants. Since 2017 it also defines a (literally 'big town') as a municipality with a population of at least 200,000 of which at least 200,000 are in its largest . This means that Stockholm, Gothenburg and Malmö are , i.e. 'major cities', while Uppsala, with a population of approximately 230,000 in the municipality, which covers an unusually large area, almost three times larger than the combined land area of the municipalities of Stockholm, Göteborg and Malmö, is not. The largest contiguous urban area within Uppsala municipality has a population of well below 200,000, while the population of both Malmö municipality, with a land area only 1/14 the size of Uppsala municipality, and Malmö , i.e. contiguous urban area, is well over 300,000, and the population of the Malmö Metropolitan Area, with a land area only slightly larger than Uppsala Municipality, is well over 700,000. A difference in the size and population of the urban area between Uppsala and the smallest in Sweden, Malmö, is the reason Statistics Sweden changed the definition of in 2017.
Ukraine
In Ukraine the term town (, ) existed from the Medieval period until 1925, when it was replaced by the Soviet regime with urban type settlement. Historically, a town in the Ukrainian lands was a smaller populated place that was chartered under the German town law and had a market square (see Market town).
Informally, towns are also referred to today as cities of district significance, cities with small population, and former Jewish shtetls.
United Kingdom
England and Wales
In England and Wales, a town traditionally was a settlement which had a charter to hold a market or fair and therefore became a "market town". Market towns were distinguished from villages in that they were the economic hub of a surrounding area, and were usually larger and had more facilities.
In parallel with popular usage, however, there are many technical and official definitions of what constitutes a town, to which various interested parties cling.
In modern official usage the term town is employed either for old market towns, or for settlements which have a town council, or for settlements which elsewhere would be classed a city, but which do not have the legal right to call themselves such. Any parish council can decide to describe itself as a town council, but this will usually only apply to the smallest "towns" (because larger towns will be larger than a single civil parish).
Not all settlements which are commonly described as towns have a town council or borough council. In fact, because of many successive changes to the structure of local government, there are now few large towns which are represented by a body closely related to their historic borough council. These days, a smaller town will usually be part of a local authority which covers several towns. And where a larger town is the seat of a local authority, the authority will usually cover a much wider area than the town itself (either a large rural hinterland, or several other, smaller towns).
Additionally, there are "new towns" which were created during the 20th century, such as Basildon, Redditch and Telford. Milton Keynes was designed to be a "new city" but legally it is still a town despite its size.
Some settlements which describe themselves as towns (e.g. Shipston-on-Stour, Warwickshire) are smaller than some large villages (e.g. Kidlington, Oxfordshire).
The status of a city is reserved for places that have letters patent entitling them to the name, historically associated with the possession of a cathedral. Some large municipalities (such as Northampton and Bournemouth) are legally boroughs but not cities, whereas some cities are quite small — such as Ely or St David's. The city of Brighton and Hove was created from the two former towns and some surrounding villages, and within the city the correct term for the former distinct entities is somewhat unclear.
It appears that a city may become a town, though perhaps only through administrative error: Rochester in Kent had been a city for centuries but, when in 1998 the Medway district was created, a bureaucratic blunder meant that Rochester lost its official city status and is now technically a town.
It is often thought that towns with bishops' seats rank automatically as cities: however, Chelmsford was a town until 5 June 2012 despite being the seat of the diocese of Chelmsford, created in 1914. St Asaph, which is the seat of the diocese of St Asaph, only became a city on 1 June 2012 though the diocese was founded in the mid sixth century. In reality, the pre-qualification of having a cathedral of the established Church of England, and the formerly established Church in Wales or Church of Ireland, ceased to apply from 1888.
The word town can also be used as a general term for urban areas, including cities and in a few cases, districts within cities. In this usage, a city is a type of town; a large one, with a certain status. For example, central Greater London is sometimes referred to colloquially as "London town". (The "City of London" is the historical nucleus, informally known as the "Square Mile", and is administratively separate from the rest of Greater London, while the City of Westminster is also technically a city and is also a London borough.) Camden Town and Somers Town are districts of London, as New Town is a district of Edinburgh – actually the Georgian centre.
In recent years the division between cities and towns has grown, leading to the establishment of groups like the Centre for Towns, who work to highlight the issues facing many towns. Towns also became a significant issue in the 2020 Labour Party leadership election, with Lisa Nandy making significant reference to Labour needing to win back smaller towns which have swung away from the party.
Scotland
A town in Scotland has no specific legal meaning and (especially in areas which were or are still Gaelic-speaking) can refer to a mere collection of buildings (e.g. a farm-town or in Scots ), not all of which might be inhabited, or to an inhabited area of any size which is not otherwise described in terms such as city, burgh, etc. Many locations of greatly different size will be encountered with a name ending with -town, -ton, -toun etc. (or beginning with the Gaelic equivalent etc.).
A burgh (pronounced burruh) is the Scots' term for a town or a municipality. They were highly autonomous units of local government from at least the 12th century until their abolition in 1975, when a new regional structure of local government was introduced across the country. Usually based upon a town, they had a municipal corporation and certain rights, such as a degree of self-governance and representation in the sovereign Parliament of Scotland adjourned in 1707.
The term no longer describes units of local government although various claims are made from time to time that the legislation used was not competent to change the status of the Royal Burghs described below. The status is now chiefly ceremonial but various functions have been inherited by current Councils (e.g. the application of various endowments providing for public benefit) which might only apply within the area previously served by a burgh; in consequence a burgh can still exist (if only as a defined geographical area) and might still be signed as such by the current local authority. The word 'burgh' is generally not used as a synonym for 'town' or 'city' in everyday speech, but is reserved mostly for government and administrative purposes.
Historically, the most important burghs were royal burghs, followed by burghs of regality and burghs of barony. Some newer settlements were only designated as police burghs from the 19th century onward, a classification which also applies to most of the older burghs.
United States
The definition of town varies widely from state to state and in many states there is no official definition. In some states, the term town refers to an area of population distinct from others in some meaningful dimension, typically population or type of government. The characteristic that distinguishes a town from another type of populated place — a city, borough, village, or township, for example — differs from state to state. In some states, a town is an incorporated municipality; that is, one with a charter received from the state, similar to a city (see incorporated town), while in others, a town is unincorporated. In some instances, the term town refers to a small incorporated municipality of less than a population threshold specified by state statute, while in others a town can be significantly larger. Some states do not use the term town at all, while in others the term has no official meaning and is used informally to refer to a populated place, of any size, whether incorporated or unincorporated. In still other states, the words town and city are legally interchangeable.
Small-town life has been a major theme in American literature, especially stories of rejection by young people leaving for the metropolis.
Since the use of the term varies considerably by state, individual usages are presented in the following sections:
Alabama
In Alabama, the legal use of the terms town and city is based on population. A municipality with a population of 2,000 or more is a city, while less than 2,000 is a town (Code of Alabama 1975, Section 11-40-6). For legislative purposes, municipalities are divided into eight classes based on population. Class 8 includes all towns, plus cities with populations of less than 6,000 (Code of Alabama 1975, Section 11-40-12).
Arizona
In Arizona, the terms town and city are largely interchangeable. A community may incorporate under either a town or a city organization with no regard to population or other restrictions according to Arizona law (see Arizona Revised Statutes, Title 9). Cities may function under slightly differing governmental systems, such as the option to organize a district system for city governments, but largely retain the same powers as towns. Arizona law also allows for the consolidation of neighboring towns and the unification of a city and a town, but makes no provision for the joining of two adjacent cities.
California
In California, the words town and city are synonymous by law (see Cal. Govt. Code Secs. 34500–34504). There are two types of cities in California: charter and general law. Cities organized as charter cities derive their authority from a charter that they draft and file with the state, and which, among other things, states the municipality's name as "City of (Name)" or "Town of (Name)." Government Code Sections 34500–34504 applies to cities organized as general law cities, which differ from charter cities in that they do not have charters but instead operate with the powers conferred them by the pertinent sections of the Government Code. Like charter cities, general law cities may incorporate as "City of (Name)" or "Town of (Name)."
Some cities change what they are referred to as. The sign in front of the municipal offices in Los Gatos, California, for example, reads "City of Los Gatos", but the words engraved on the building above the front entrance when the city hall was built read "Town of Los Gatos." There are also signs at the municipal corporation limit, some of which welcome visitors to the "City of Los Gatos" while older, adjacent signs welcome people to the "Town of Los Gatos." Meanwhile, the village does not exist in California as a municipal corporation. Instead, the word town is commonly used to indicate any unincorporated community that might otherwise be known as an unincorporated village. Additionally, some people may still use the word town as shorthand for township, which is not an incorporated municipality but an administrative division of a county.
Hawaii
The Hawaiian Island of Oahu has various communities that may be referred to as towns. However, the entire island is lumped as a single incorporated city, the City and County of Honolulu. The towns on Oahu are merely unincorporated census-designated places.
Illinois
In Illinois, the word town has been used both to denote a subdivision of a county called a township, and to denote a form of municipality similar to a village, in that it is generally governed by a president and trustees rather than a mayor. In some areas a town may be incorporated legally as a village (meaning it has at large trustees) or a city (meaning it has aldermen from districts) and absorb the duties of the township it is coterminous with (maintenance of birth records, certain welfare items). Evanston, Berwyn and Cicero are examples of towns in this manner. Under the current Illinois Municipal Code, an incorporated or unincorporated town may choose to incorporate as a city or as a village, but other forms of incorporation are no longer allowed.
Louisiana
In Louisiana a town is defined as being a municipal government having a population of 1,001 to 4,999 inhabitants.
Maryland
While a town is generally considered a smaller entity than a city, the two terms are legally interchangeable in Maryland. The only exception may be the independent city of Baltimore, which is a special case, as it was created by the Constitution of Maryland.
Nevada
In Nevada, a town has a form of government, but is not considered to be incorporated. It generally provides a limited range of services, such as land use planning and recreation, while leaving most services to the county. Many communities have found this "semi-incorporated" status attractive; the state has only 20 incorporated cities, and towns as large as Paradise (186,020 in 2000 Census), home of the Las Vegas Strip. Most county seats are also towns, not cities.
New England
In the six New England states, a town is the most prevalent minor civil division, and in most cases, are a more important form of government than the county. In Connecticut, Rhode Island and seven out of fourteen counties in Massachusetts, in fact, counties only exist as boundaries for state services and chambers of commerce at most, and have no independent legal functions. In New Hampshire, Maine, and Vermont, counties function at a limited scope, and are still not as important in northern New England as they are outside of the northeast. In all six states, towns perform functions that in most states would be county functions. The defining feature of a New England town, as opposed to a city, is that a town meeting and a board of selectmen serve as the main form of government for a town, while cities are run by a mayor and a city council. For example, Brookline, Massachusetts is a town, even though it is fairly urban, because of its form of government. In the three southern New England states, the entire land area is divided into towns and cities, while the three northern states have small areas that are unincorporated. In Vermont and New Hampshire, the population of these areas is practically nonexistent, while in Maine, unincorporated areas make up roughly half of the state’s area but only one percent of the state’s population.
Though the U.S. Census Bureau defines New England towns as "minor civil divisions" for statistical purposes, all New England towns are municipal corporations equivalent to cities in all legal respects, except for form of government. For statistical purposes, the Census Bureau uses census-designated places for the built-up population centers within towns, though these have no legal or social recognition for residents of those towns. Similarly, the Census Bureau uses a special designation for urban areas within New England, the New England city and town area, instead of the metropolitan statistical area it uses in the rest of the country.
New Jersey
A town in the context of New Jersey local government refers to one of five types and one of eleven forms of municipal government. While town is often used as a shorthand to refer to a township, the two are not the same. The Town Act of 1895 allowed any municipality or area with a population exceeding 5,000 to become a Town through a petition and referendum process. Under the 1895 Act, a newly incorporated town was divided into at least three wards, with two councilmen per ward serving staggered two-year terms, and one councilman at large, who also served a two-year term. The councilman at large served as chairman of the town council. The Town Act of 1988 completely revised the town form of government and applied to all towns incorporated under the Town Act of 1895 and to those incorporated by a special charter granted by the Legislature prior to 1875.
Under the 1988 Act, the mayor is also the councilman at large, serving a term of two years, unless increased to three years by a petition and referendum process. The council under the Town Act of 1988 consists of eight members serving staggered two-year terms with two elected from each of four wards. One council member from each ward is up for election each year. Towns with different structures predating the 1988 Act may retain those features unless changed by a petition and referendum process. Two new provisions were added in 1991 to the statutes governing towns, First, a petition and referendum process was created whereby the voters can require that the mayor and town council be elected to four-year terms of office. The second new provision defines the election procedure in towns with wards. The mayor in a town chairs the town council and heads the municipal government. The mayor may both vote on legislation before council and veto ordinances. A veto may be overridden by a vote of two-thirds of all the members of the council. The council may enact an ordinance to delegate all or a portion of the executive responsibilities of the town to a municipal administrator. Fifteen New Jersey municipalities currently have a type of town, nine of which operate under the town form of government.
New York
In New York, a town is a division of the county that possesses home rule powers, but generally with less functions than towns in New England. A town provides a closer level of governance than its enclosing county, providing almost all municipal services to unincorporated communities, called hamlets, and selected services to incorporated areas, called villages. In New York, a town typically contains a number of such hamlets and villages. However, due to their independent nature, incorporated villages may exist in two towns or even two counties (example: Almond (village), New York). Everyone in New York who does not live in a city or Indian reservation lives in a town and possibly in one of the town's hamlets or villages. New York City and Geneva are the only two cities that span county boundaries. The only part of Geneva in Seneca County is water; each of the boroughs of New York City is a county.
North Carolina
In North Carolina, all cities, towns, and villages are incorporated as municipalities. According to the North Carolina League of Municipalities, there is no legal distinction among a city, town, or village—it is a matter of preference of the local government. Some North Carolina cities have populations as small as 1,000 residents, while some towns, such as Cary, have populations of greater than 100,000.
Oklahoma
In Oklahoma, according to the state's municipal code, city means a municipality which has incorporated as a city in accordance with the laws of the state, whereas town means a municipality which has incorporated as a town in accordance with the laws of the state, and municipality means any incorporated city or town. The term village is not defined or used in the act. Any community of people residing in compact form may become incorporated as a town; however, if the resident population is one thousand or more, a town or community of people residing in compact form may become incorporated as a city.
Pennsylvania
In Pennsylvania, the incorporated divisions are townships, boroughs, and cities, of which boroughs are equivalent to towns (example: State College is a borough). However, one borough is incorporated as a town: Bloomsburg.
Texas
In Texas, although some municipalities refer to themselves as "towns" or "villages" (to market themselves as an attractive place to live), these names have no specific designation in Texas law; legally all incorporated places are considered cities.
Utah
In Utah, the legal use of the terms town and city is based on population. A municipality with a population of 1,000 or more is a city, while less than 1,000 is a town. In addition, cities are divided into five separate classes based on the population.
Virginia
In Virginia, a town is an incorporated municipality similar to a city (though with a smaller required minimum population). But while cities are by Virginia law independent of counties, towns are contained within counties.
Washington
A town in the state of Washington is a municipality that has a population of less than 1,500 at incorporation, however an existing town can reorganize as a code city. Town government authority is limited relative to cities, the other main classification of municipalities in the state. , most municipalities in Washington are cities. (See List of towns in Washington.)
Wisconsin
Wisconsin has towns which are areas outside of incorporated cities and villages. These towns retain the name of the civil township from which they evolved and are often the same name as a neighboring city. Some towns, especially those in urban areas, have services similar to those of incorporated cities, such as police departments. These towns will, from time to time, incorporate into cities, such as Fox Crossing in 2016 from the former town of Menasha. Often this is to protect against being annexed into neighboring cities and villages.
Wyoming
A Wyoming statute indicates towns are incorporated municipalities with populations of less than 4,000. Municipalities of 4,000 or more residents are considered "first-class cities".
Vietnam
In Vietnam, a district-level town () is the second subdivision, below a province () or municipality (). A commune-level town () a third-level (commune-level) subdivision, below a district ().
See also
Commuter town
Company town
Developed environments
List of towns
Location (geography)
Megalopolis (city type)
Proto-city
Town charter
Town Hall
Town limits
Town privileges
Town square
References
Sources
External links
Australian Bureau of Statistics: Australian Standard Geographical Classification (ASGC) 2005
Open-Site Regional — Contains information about towns in numerous countries.
Geopolis : research group, university of Paris-Diderot, France — Access to Geopolis Database
Types of populated places |
null | null | Stefan–Boltzmann constant | eng_Latn | The Stefan–Boltzmann constant (also Stefan's constant), a physical constant denoted by the Greek letter σ (sigma), is the constant of proportionality in the Stefan–Boltzmann law: "the total intensity radiated over all wavelengths increases as the temperature increases", of a black body which is proportional to the fourth power of the thermodynamic temperature. The theory of thermal radiation lays down the theory of quantum mechanics, by using physics to relate to molecular, atomic and sub-atomic levels. Slovenian physicist Josef Stefan formulated the constant in 1879; it was formally derived in 1884 by his former student and collaborator, the Austrian physicist Ludwig Boltzmann. The equation can also be derived from Planck's law, by integrating over all wavelengths at a given temperature, which will represent a small flat black body box. "The amount of thermal radiation emitted increases quickly and the principal frequency of the radiation becomes higher with increasing temperatures". The Stefan–Boltzmann constant can be used to measure the amount of heat that is emitted by a black body, which absorbs all of the radiant energy that hits it, and will emit all the radiant energy. Furthermore, the Stefan–Boltzmann constant allows for temperature (K) to be converted to units for intensity (W⋅m−2), which is power per unit area.
Value
Since the 2019 redefinition of the SI base units, the Stefan–Boltzmann constant is given exactly rather than in experimental values. The value is given in SI units by
In cgs units the Stefan–Boltzmann constant is
In thermochemistry the Stefan–Boltzmann constant is often expressed in cal⋅cm−2⋅day−1⋅K−4:
In US customary units the Stefan–Boltzmann constant is
The Stefan–Boltzmann constant is defined in terms of other fundamental constants as
where
kB is the Boltzmann constant,
h is the Planck constant,
ħ is the reduced Planck constant, and
c is the speed of light in vacuum,
giving
The CODATA recommended value [ref?] prior to 20 May 2019 (2018 CODATA) was calculated from the measured value of the gas constant:
where
R is the universal gas constant
NA is the Avogadro constant
R∞ is the Rydberg constant
Ar(e) is the "relative atomic mass" of the electron
Mu is the molar mass constant (1 g/mol by definition)
α is the fine-structure constant.
Dimensional formula: M1T−3Θ−4
A related constant is the radiation constant (or radiation density constant) , which is given by
References
External links
Physical constants
Thermodynamics
Ludwig Boltzmann
de:Stefan-Boltzmann-Gesetz |
null | null | Lord Protector | eng_Latn | Lord Protector (plural: Lords Protector) was a title that has been used in British constitutional law for the head of state. It was also a particular title for the British heads of state in respect to the established church. It was sometimes used to refer to holders of other temporary posts; for example, a regent acting for the absent monarch.
Feudal Royal Regent
The title of "The Lord Protector" was originally used by royal princes or other nobles exercising an individual regency (i.e., not merely as a member of a collegial regency council) while the English monarch was still a minor or otherwise unable to rule.
Notable cases in England:
John, Duke of Bedford, and Humphrey, Duke of Gloucester, were (5 December 1422 – 6 November 1429) jointly Lords Protector for Henry VI (1421–1471);
Richard Plantagenet, Duke of York, was thrice (3 April 1454 – February 1455; 19 November 1455 – 25 February 1456; and 31 October – 30 December 1460) Lord Protector for Henry VI;
Richard, Duke of Gloucester, was "Lord Protector of the Realm" (10 May 1483 – 26 June 1483) during the nominal reign of Edward V (one of the Princes in the Tower) before claiming the throne for himself as Richard III;
Edward Seymour, Duke of Somerset, was Lord Protector (4 February 1547 – 11 October 1549) during the early years of the reign of the young Edward VI;
and in Scotland:
John Stewart, Duke of Albany, was "Governor and Protector of the Realm" (12 July 1515 – 16 November 1524) for James V of Scotland (1512–1542);
James Hamilton, Duke of Châtellerault, was "Governor and Protector of the Kingdom" (3 January 1543 – 12 April 1554) for Mary, Queen of Scots.
Cromwellian Commonwealth
The Lord Protector of the Commonwealth of England, Scotland and Ireland was the title of the head of state during the Commonwealth (often called the Interregnum), following the first period when a Council of State held executive power. The title was held by Oliver Cromwell (December 1653 – September 1658) and subsequently his son and designated successor Richard Cromwell (September 1658 – May 1659) during what is now known as The Protectorate.
The 1653 Instrument of Government (republican constitution) stated:
The replacement constitution of 1657, the Humble Petition and Advice, gave "His Highness the Lord Protector" the power to nominate his successor. Cromwell chose his eldest surviving son, the politically inexperienced Richard. That was a nonrepresentative and de facto dynastic mode of succession, with royal connotations in both styles awarded, (even a double invocation 16 December 1653 – 3 September 1658 "By the Grace of God and Republic Lord Protector of England, Scotland and Ireland") and many other monarchic prerogatives, such as that of awarding knighthoods.
Richard Cromwell, who succeeded after his father's death in September 1658, held the position for only eight months before he resigned in May 1659. He was followed by the second period of Commonwealth rule until the Restoration of the exiled heir to the Stuart throne, Charles II, in May 1660.
Lords Protector (1653–59)
Post-Cromwell
Since the Restoration the title has not been used in either of the above manners. George, Prince of Wales, appointed to the regency in 1811, was referred to as "His Royal Highness the Prince Regent". George exercised the powers of the monarchy, just as Lords Protector had, but the title's republican associations had rendered it distasteful.
Protector of the church
'Lord Protector' has also been used as a rendering of the Latin Advocatus in the sense of a temporal Lord (such as a Monarch) who acted as the protector of the mainly secular interests of a part of the church; compare the French title of vidame.
References
External links
United Kingdom at WorldStatesmen.org
Heads of state of England
Heads of state
Lords Protector of England
Republicanism in the United Kingdom
Stuart England
Interregnum (England) |
null | null | Number of the Beast | eng_Latn | The number of the beast, 666 in most manuscripts, is associated with the Beast of Revelation in the Book of Revelation in the Bible.
Number of the Beast also may refer to:
The Number of the Beast (album), 1982, by Iron Maiden
"The Number of the Beast" (song), from the album
The Number of the Beast (play), a 1982 drama by Snoo Wilson about Aleister Crowley
Number of the Beast (comics), a comic book limited series
The Number of the Beast (novel), by Robert Heinlein
"The Number of the Beast Is 666" (Hannibal), an episode of the TV series
The Number of the Beast is 666, one of The Great Red Dragon paintings by William Blake
See also
Mark of the Beast (disambiguation)
616 (number)#Number of the beast, another interpretation of the number of the beast |
null | null | Vitamin E | eng_Latn | Vitamin E is a group of eight fat soluble compounds that include four tocopherols and four tocotrienols. Vitamin E deficiency, which is rare and usually due to an underlying problem with digesting dietary fat rather than from a diet low in vitamin E, can cause nerve problems. Vitamin E is a fat-soluble antioxidant which may help protect cell membranes from reactive oxygen species.
Worldwide, government organizations recommend adults consume in the range of 3 to 15 mg per day. As of 2016, consumption was below recommendations according to a worldwide summary of more than one hundred studies that reported a median dietary intake of 6.2 mg per day for alpha-tocopherol. Research with alpha-tocopherol as a dietary supplement, with daily amounts as high as 2,000 mg per day, has had mixed results. Population studies suggested that people who consumed foods with more vitamin E, or who chose on their own to consume a vitamin E dietary supplement, had lower incidence of cardiovascular diseases, cancer, dementia, and other diseases, but placebo-controlled clinical trials could not always replicate these findings. As of 2017, vitamin E continues to be a topic of active clinical research. There is no clinical evidence that use of vitamin E skincare products are effective. Both natural and synthetic tocopherols are subject to oxidation, and so in dietary supplements are esterified, creating tocopheryl acetate for stability purposes.
Both the tocopherols and tocotrienols occur in α (alpha), β (beta), γ (gamma) and δ (delta) forms, as determined by the number and position of methyl groups on the chromanol ring. All eight of these vitamers feature a chromane double ring, with a hydroxyl group that can donate a hydrogen atom to reduce free radicals, and a hydrophobic side chain which allows for penetration into biological membranes.
Vitamin E was discovered in 1922, isolated in 1935 and first synthesized in 1938. Because the vitamin activity was first identified as essential for fertilized eggs to result in live births (in rats), it was given the name "tocopherol" from Greek words meaning birth and to bear or carry. Alpha-tocopherol, either naturally extracted from plant oils or, most commonly, as the synthetic tocopheryl acetate, is sold as a popular dietary supplement, either by itself or incorporated into a multivitamin product, and in oils or lotions for use on skin.
Chemistry
The nutritional content of vitamin E is defined by equivalency to 100% RRR-configuration α-tocopherol activity. The molecules that contribute α-tocopherol activity are four tocopherols and four tocotrienols, within each group of four identified by the prefixes alpha- (α-), beta- (β-), gamma- (γ-), and delta- (δ-). For alpha(α)-tocopherol each of the three "R" sites has a methyl group (CH3) attached. For beta(β)-tocopherol: R1 = methyl group, R2 = H, R3 = methyl group. For gamma(γ)-tocopherol: R1 = H, R2 = methyl group, R3 = methyl group. For delta(δ)-tocopherol: R1 = H, R2 = H, R3 = methyl group. The same configurations exist for the tocotrienols, except that the hydrophobic side chain has three carbon-carbon double bonds whereas the tocopherols have a saturated side chain.
Stereoisomers
In addition to distinguishing tocopherols and tocotrienols by position of methyl groups, the tocopherols have a phytyl tail with three chiral points or centers that can have a right or left orientation. The naturally occurring plant form of alpha-tocopherol is RRR-α-tocopherol, also referred to as d-tocopherol, whereas the synthetic form (all-racemic or all-rac vitamin E, also dl-tocopherol) is equal parts of eight stereoisomers RRR, RRS, RSS, SSS, RSR, SRS, SRR and SSR with progressively decreasing biological equivalency, so that 1.36 mg of dl-tocopherol is considered equivalent to 1.0 mg of d-tocopherol, the natural form. Rephrased, the synthetic has 73.5% of the potency of the natural.
Tocopherols
Alpha-tocopherol is a lipid-soluble antioxidant functioning within the glutathione peroxidase pathway, and protecting cell membranes from oxidation by reacting with lipid radicals produced in the lipid peroxidation chain reaction. This removes the free radical intermediates and prevents the oxidation reaction from continuing. The oxidized α-tocopheroxyl radicals produced in this process may be recycled back to the active reduced form through reduction by other antioxidants, such as ascorbate, retinol or ubiquinol. Other forms of vitamin E have their own unique properties; for example, γ-tocopherol is a nucleophile that can react with electrophilic mutagens.
Tocotrienols
The four tocotrienols (alpha, beta, gamma, delta) are similar in structure to the four tocopherols, with the main difference being that the former have hydrophobic side chains with three carbon-carbon double bonds, whereas the tocopherols have saturated side chains. For alpha(α)-tocotrienol each of the three "R" sites has a methyl group (CH3) attached. For beta(β)-tocotrienol: R1 = methyl group, R2 = H, R3 = methyl group. For gamma(γ)-tocotrienol: R1 = H, R2 = methyl group, R3 = methyl group. For delta(δ)-tocotrienol: R1 = H, R2 = H, R3 = methyl group. Palm oil is a good source of alpha and gamma tocotrienols.
Tocotrienols have only a single chiral center, which exists at the 2' chromanol ring carbon, at the point where the isoprenoid tail joins the ring. The other two corresponding centers in the phytyl tail of the corresponding tocopherols do not exist as chiral centers for tocotrienols due to unsaturation (C-C double bonds) at these sites. Tocotrienols extracted from plants are always dextrorotatory stereoisomers, signified as d-tocotrienols. In theory, levorotatory forms of tocotrienols (l-tocotrienols) could exist as well, which would have a 2S rather than 2R configuration at the molecules' single chiral center, but unlike synthetic dl-alpha-tocopherol, the marketed tocotrienol dietary supplements are all d-tocotrienol extracts from palm or annatto oils. Preliminary clinical trials on dietary supplement tocotrienols indicate potential for anti-disease activity.
Functions
Vitamin E may have various roles as a vitamin. Many biological functions have been postulated, including a role as a fat-soluble antioxidant. In this role, vitamin E acts as a radical scavenger, delivering a hydrogen (H) atom to free radicals. At 323 kJ/mol, the O-H bond in tocopherols is about 10% weaker than in most other phenols. This weak bond allows the vitamin to donate a hydrogen atom to the peroxyl radical and other free radicals, minimizing their damaging effect. The thus-generated tocopheryl radical is recycled to tocopherol by a redox reaction with a hydrogen donor, such as vitamin C. As it is fat-soluble, vitamin E is incorporated into cell membranes, which are therefore protected from oxidative damage.
Vitamin E affects gene expression and is an enzyme activity regulator, such as for protein kinase C (PKC) – which plays a role in smooth muscle growth – with vitamin E participating in deactivation of PKC to inhibit smooth muscle growth.
Synthesis
Biosynthesis
Photosynthesizing plants, algae and cyanobacteria synthesize tocochromanols, the chemical family of compounds made up of four tocopherols and four tocotrienols; in a nutrition context this family is referred to as Vitamin E. Biosynthesis starts with formation of the closed-ring part of the molecule as homogentisic acid (HGA). The side chain is attached (saturated for tocopherols, polyunsaturated for tocotrienols). The pathway for both is the same, so that gamma- is created and from that alpha-, or delta- is created and from that the beta- compounds. Biosynthesis takes place in the plastids.
As to why plants synthesize tocochromanols, the major reason appears to be for antioxidant activity. Different parts of plants, and different species, are dominated by different tocochromanols. The predominant form in leaves, and hence leafy green vegetables is α-tocopherol. Location is in chloroplast membranes, in close proximity to the photosynthetic process. The function is to protect against damage from the ultraviolet radiation of sunlight. Under normal growing conditions the presence of α-tocopherol does not appear to be essential, as there are other photo-protective compounds, and plants that through mutations have lost the ability to synthesize α-tocopherol demonstrate normal growth. However, under stressed growing conditions such as drought, elevated temperature or salt-induced oxidative stress, the plants' physiological status is superior if it has the normal synthesis capacity.
Seeds are lipid-rich, to provide energy for germination and early growth. Tocochromanols protect the seed lipids from oxidizing and becoming rancid. The presence of tocochromanols extends seed longevity, and promotes successful germination and seedling growth. Gamma-tocopherol dominates in seeds of most plant species, but there are exceptions. For canola, corn and soy bean oils, there is more γ-tocopherol than α-tocopherol, but for safflower, sunflower and olive oils the reverse is true. Of the commonly used food oils, palm oil is unique in that tocotrienol content is higher than tocopherol content. Seed tocochromanols content is also dependent on environmental stressors. In almonds, for example, drought or elevated temperature increase α-tocopherol and γ-tocopherol content of the nuts. The same article mentions that drought increases the tocopherol content of olives, and heat likewise for soybeans.
Industrial synthesis
Naturally sourced d-alpha-tocopherol can be extracted and purified from seed oils, or gamma-tocopherol can be extracted, purified, and methylated to create d-alpha-tocopherol. In contrast to alpha-tocopherol extracted from plants, which is also called d-alpha-tocopherol, industrial synthesis creates dl-alpha-tocopherol. "It is synthesized from a mixture of toluene and 2,3,5-trimethyl-hydroquinone that reacts with isophytol to all-rac-alpha-tocopherol, using iron in the presence of hydrogen chloride gas as catalyst. The reaction mixture obtained is filtered and extracted with aqueous caustic soda. Toluene is removed by evaporation and the residue (all rac-alpha-tocopherol) is purified by vacuum distillation." Specification for the ingredient is >97% pure. This synthetic dl-alpha-tocopherol has approximately 50% of the potency of d-alpha-tocopherol. Manufacturers of dietary supplements and fortified foods for humans or domesticated animals convert the phenol form of the vitamin to an ester using either acetic acid or succinic acid because the esters are more chemically stable, providing for a longer shelf-life. The ester forms are de-esterified in the gut and absorbed as free alpha-tocopherol.
Deficiency
Vitamin E deficiency is rare in humans, occurring as a consequence of abnormalities in dietary fat absorption or metabolism rather than from a diet low in vitamin E. One example of a genetic abnormality in metabolism is mutations of genes coding for alpha-tocopherol transfer protein (α-TTP). Humans with this genetic defect exhibit a progressive neurodegenerative disorder known as ataxia with vitamin E deficiency (AVED) despite consuming normal amounts of vitamin E. Large amounts of alpha-tocopherol as a dietary supplement are needed to compensate for the lack of α-TTP Vitamin E deficiency due to either malabsorption or metabolic anomaly can cause nerve problems due to poor conduction of electrical impulses along nerves due to changes in nerve membrane structure and function. In addition to ataxia, vitamin E deficiency can cause peripheral neuropathy, myopathies, retinopathy and impairment of immune responses.
Drug interactions
The amounts of alpha-tocopherol, other tocopherols and tocotrienols that are components of dietary vitamin E, when consumed from foods, do not appear to cause any interactions with drugs. Consumption of alpha-tocopherol as a dietary supplement in amounts in excess of 300 mg/day may lead to interactions with aspirin, warfarin, tamoxifen and cyclosporine A in ways that alter function. For aspirin and warfarin, high amounts of vitamin E may potentiate anti-blood clotting action. One small trial demonstrated that vitamin E at 400 mg/day reduced blood concentration of the anti-breast cancer drug tamoxifen. In multiple clinical trials, vitamin E lowered blood concentration of the immunosuppressant medication, cyclosporine A. The US National Institutes of Health, Office of Dietary Supplements, raises a concern that co-administration of vitamin E could counter the mechanisms of anti-cancer radiation therapy and some types of chemotherapy, and so advises against its use in these patient populations. The references it cited reported instances of reduced treatment adverse effects, but also poorer cancer survival, raising the possibility of tumor protection from the intended oxidative damage by the treatments.
Dietary recommendations
The U.S. National Academy of Medicine updated estimated average requirements (EARs) and recommended dietary allowances (RDAs) for vitamin E in 2000. RDAs are higher than EARs so as to identify amounts that will cover people with higher than average requirements. Adequate intakes (AIs) are identified when there is not sufficient information to set EARs and RDAs. The EAR for vitamin E for women and men ages 14 and up is 12 mg/day. The RDA is 15 mg/day. As for safety, tolerable upper intake levels ("upper limits" or ULs) are set for vitamins and minerals when evidence is sufficient. Hemorrhagic effects in rats were selected as the critical endpoint to calculate the upper limit via starting with the lowest-observed-adverse-effect-level. The result was a human upper limit set at 1000 mg/day. Collectively the EARs, RDAs, AIs and ULs are referred to as Dietary Reference Intakes.
The European Food Safety Authority (EFSA) refers to the collective set of information as dietary reference values, with population reference intakes (PRIs) instead of RDAs, and average requirements instead of EARs. AIs and ULs are defined the same as in the United States. For women and men ages 10 and older, the PRIs are set at 11 and 13 mg/day, respectively. PRI for pregnancy is 11 mg/day, for lactation 11 mg/day. For children ages 1–9 years the PRIs increase with age from 6 to 9 mg/day. The EFSA used an effect on blood clotting as a safety-critical effect. It identified that no adverse effects were observed in a human trial as 540 mg/day, used an uncertainty factor of 2 to get to a suggest an upper limit of half of that, then rounded to 300 mg/day.
The Japan National Institute of Health and Nutrition set adult AIs at 6.5 mg/day (females) and 7.0 mg/day (males), and 650–700 mg/day (females), and 750–900 mg/day (males) for upper limits, amounts depending on age. India recommends an intake of 8–10 mg/day and does not set an upper limit. The World Health Organization recommends that adults consume 10 mg/day. The United Kingdom is an outlier, in that it recommends 4 mg/day for adult men and 3 mg/day for adult women.
Consumption is below these government recommendations. Government survey results in the United States reported average consumption for adult females at 8.4 mg/d and adult males 10.4 mg/d. Both are below the RDA of 15 mg/day.A worldwide summary of more than one hundred studies reported a median dietary intake of 6.2 mg/d for alpha-tocopherol.
Food labeling
For U.S. food and dietary supplement labeling purposes the amount in a serving is expressed as a percent of daily value. For vitamin E labeling purposes 100% of the daily value was 30 international units, but as of 27 May 2016 it was revised to 15 mg to bring it into agreement with the RDA. Compliance with the updated labeling regulations was required by 1 January 2020 for manufacturers with US$10 million or more in annual food sales, and by 1 January 2021 for manufacturers with lower volume food sales. A table of the old and new adult daily values is provided at Reference Daily Intake.
European Union regulations require that labels declare energy, protein, fat, saturated fat, carbohydrates, sugars, and salt. Voluntary nutrients may be shown if present in significant amounts. Instead of daily values, amounts are shown as percent of reference intakes (RIs). For vitamin E, 100% RI was set at 12 mg in 2011.
The international unit measurement was used by the United States in 1968–2016. 1 IU is the biological equivalent of about 0.667 mg d (RRR)-alpha-tocopherol (2/3 mg exactly), or of 0.90 mg of dl-alpha-tocopherol, corresponding to the then-measured relative potency of stereoisomers. In May 2016, the measurements have been revised, such that 1 mg of "Vitamin E" is 1 mg of d-alpha-tocopherol or 2 mg of dl-alpha-tocopherol. The change was originally started in 2000, when forms of Vitamin E other than alpha-tocopherol was dropped from dietary calculations by the IOM. The UL amount disregards any conversion. The EFSA has never used an IU unit, and their measurement only considers RRR-alpha-tocopherol.
Sources
Worldwide, consumption is below recommendations according to a summary of more than one hundred studies that reported a median dietary intake of 6.2 mg per day for alpha-tocopherol. Of the many different forms of vitamin E, gamma-tocopherol (γ-tocopherol) is the most common form found in the North American diet, but alpha-tocopherol (α-tocopherol) is the most biologically active. Palm oil is a source of tocotrienols.
The U.S. Department of Agriculture (USDA), Agricultural Research Services, maintains a food composition database. The last major revision was Release 28, September 2015. In addition to the naturally occurring sources shown in the table, certain ready-to-eat cereals, infant formulas, liquid nutrition products and other foods are fortified with alpha-tocopherol.
Supplements
Vitamin E is fat soluble, so dietary supplement products are usually in the form of the vitamin, esterified with acetic acid to generate tocopheryl acetate, and dissolved in vegetable oil in a softgel capsule. For alpha-tocopherol, amounts range from 100 to 1000 IU per serving. Smaller amounts are incorporated into multi-vitamin/mineral tablets. Gamma-tocopherol and tocotrienol supplements are also available from dietary supplement companies. The latter are extracts from palm or annatto oils.
Fortification
The World Health Organization does not have any recommendations for food fortification with vitamin E. The Food Fortification Initiative does not list any countries that have mandatory or voluntary programs for vitamin E. Infant formulas have alpha-tocopherol as an ingredient. In some countries, certain brands of ready-to-eat cereals, liquid nutrition products and other foods have alpha-tocopherol as an added ingredient.
Food additives
Various forms of vitamin E are common food additive in oily food, used to deter rancidity caused by peroxidation. Those with an E number include:
E306 Tocopherol-rich extract (mixed, natural, can include tocotrienol)
E307 Alpha-tocopherol (synthetic)
E308 Gamma-tocopherol (synthetic)
E309 Delta-tocopherol (synthetic)
These E numbers include all racemic forms and acetate esters thereof. Commonly found on food labels in Europe and some other countries, their safety assessment and approval are the responsibility of the European Food Safety Authority.
Metabolism
Tocotrienols and tocopherols, the latter including the stereoisomers of synthetic alpha-tocopherol, are absorbed from the intestinal lumen, incorporated into chylomicrons, and secreted into the portal vein, leading to the liver. Absorption efficiency is estimated at 51% to 86%, and that applies to all of the vitamin E family – there is no discrimination among the vitamin E vitamers during absorption. Unabsorbed vitamin E is excreted via feces. Additionally, vitamin E is excreted by the liver via bile into the intestinal lumen, where it will either be reabsorbed or excreted via feces, and all of the vitamin E vitamers are metabolized and then excreted via urine.
Upon reaching the liver, RRR-alpha-tocopherol is preferentially taken up by alpha-tocopherol transfer protein (α-TTP). All other forms are degraded to 2'-carboxethyl-6-hydroxychromane (CEHC), a process that involves truncating the phytic tail of the molecule, then either sulfated or glycuronidated. This renders the molecules water-soluble and leads to excretion via urine. Alpha-tocopherol is also degraded by the same process, to 2,5,7,8-tetramethyl-2-(2'-carboxyethyl)-6-hydroxychromane (α-CEHC), but more slowly because it is partially protected by α-TTP. Large intakes of α-tocopherol result in increased urinary α-CEHC, so this appears to be a means of disposing of excess vitamin E.
Alpha-tocopherol transfer protein is coded by the TTPA gene on chromosome 8. The binding site for RRR-α-tocopherol is a hydrophobic pocket with a lower affinity for beta-, gamma-, or delta-tocopherols, or for the stereoisomers with an S configuration at the chiral 2 site. Tocotrienols are also a poor fit because the double bonds in the phytic tail create a rigid configuration that is a mismatch with the α-TTP pocket. A rare genetic defect of the TTPA gene results in people exhibiting a progressive neurodegenerative disorder known as ataxia with vitamin E deficiency (AVED) despite consuming normal amounts of vitamin E. Large amounts of alpha-tocopherol as a dietary supplement are needed to compensate for the lack of α-TTP The role of α-TTP is to move α-tocopherol to the plasma membrane of hepatocytes (liver cells), where it can be incorporated into newly created very low density lipoprotein (VLDL) molecules. These convey α-tocopherol to cells in the rest of the body. As an example of a result of the preferential treatment, the US diet delivers approximately 70 mg/d of γ-tocopherol and plasma concentrations are on the order of 2–5 µmol/L; meanwhile, dietary α-tocopherol is about 7 mg/d but plasma concentrations are in the range of 11–37 µmol/L.
Affinity of α-TTP for vitamin E vitamers
Testing for levels
A worldwide summary of more than one hundred human studies reported a median of 22.1 µmol/L for serum α-tocopherol, and defined α-tocopherol deficiency as less than 12 µmol/L. It cited a recommendation that serum α-tocopherol concentration be ≥30 µmol/L to optimize health benefits. In contrast, the U.S. Dietary Reference Intake text for vitamin E concluded that a plasma concentration of 12 µmol/L was sufficient to achieve normal ex vivo hydrogen peroxide-induced hemolysis. A 2014 review defined less than 9 µmol/L as deficient, 9-12 µmol/L as marginal, and greater than 12 µmol/L as adequate.
Serum concentration increases with age. This is attributed to the fact that vitamin E circulates in blood incorporated into lipoproteins, and serum lipoprotein concentrations increase with age. Infants and young children have a higher risk of being below the deficiency threshold. Cystic fibrosis and other fat malabsorption conditions can result in low serum vitamin E. Dietary supplements will raise serum vitamin E.
Research
For the conditions described below, the results of RCTs do not always concur with the observational evidence. This could be a matter of amount. Observational studies compare low consumers to high consumers based on intake from food, whereas RCTs often used amounts of alpha-tocopherol 20X to 30X higher than what can be achieved from food. Diets higher in vitamin E may contain other compounds that convey health benefits, so the observed effect may not be due to the vitamin E content. There is also a concern that supplementing with alpha-tocopherol in multiples much higher than is possible via diet will suppress absorption and retention of other tocopherols, with unknown effects on health. Supplementing alpha-tocopherol is known to reduce serum gamma- and delta-tocopherol concentrations. From one large survey, consumption of alpha-tocopherol as a supplement lowered serum gamma-tocopherol from 6.0 micromol/L for people not consuming any supplement to 2.1 micromol/L for those consuming greater than or equal to 400 IU/day.
All-cause mortality
Two meta-analyses concluded that as a dietary supplement, vitamin E neither improved nor impaired all-cause mortality. An older meta-analysis had concluded high-dosage (≥400 IU/d for at least 1 year) vitamin E was associated with an increase in all-cause mortality. The authors acknowledged that the cited high-dose trials were often small and performed with people who already had chronic diseases. A meta-analysis of long-term clinical trials reported a non-significant 2% increase in all-cause mortality when alpha-tocopherol was the only supplement used. The same meta-analysis reported a statistically significant 3% increase for results when alpha-tocopherol was used by itself or in combination with other nutrients (vitamin A, vitamin C, beta-carotene, selenium).
Declining supplement use
In the United States vitamin E supplement use by female health professionals was 16.1% in 1986, 46.2% in 1998, 44.3% in 2002, but decreased to 19.8% in 2006. Similarly, for male health professionals, rates for same years were 18.9%, 52.0%, 49.4% and 24.5%. The authors theorized that declining use in these populations may have been due to publications of studies that showed either no benefits or negative consequences from vitamin E supplements. Within the U.S. military services, vitamin prescriptions written for active, reserve and retired military, and their dependents, were tracked over years 2007–2011. Vitamin E prescriptions decreased by 53% while vitamin C remained constant and vitamin D increased by 454%. A report on vitamin E sales volume in the US documented a 50% decrease between 2000 and 2006, with a potential reason being a meta-analysis that concluded high-dosage (≥400 IU/d for at least 1 year) vitamin E was associated with an increase in all-cause mortality.
Age-related macular degeneration
A Cochrane review reported no change to risk of developing age-related macular degeneration from long-term vitamin E suplimentation.
Alzheimer's disease
A 2012 review of dietary intake studies reported that higher consumption of vitamin E from foods lowered the risk of developing Alzheimer's disease (AD) by 24%. In 2017, a consensus statement from the British Association for Psychopharmacology concluded that, until further information is available, vitamin E cannot be recommended for treatment or prevention of Alzheimer's disease. A 2017 Cochrane review reported on vitamin E as a potential dietary benefit for mild cognitive impairment (MCI) and Alzheimer's disease. Based on evidence from one trial in each of the categories, the study found insufficient evidence for supplemental vitamin E to prevent progression from MCI to dementia, but it did indicate slowing of functional decline in people with AD. Given the small number of trials and subjects, the authors recommended further research. A 2018 review found lower vitamin E blood levels in AD people compared to healthy, age-matched people.
Cancer
An inverse relationship between dietary vitamin E and kidney cancer and bladder cancer is seen in observational studies. The risk reduction was 19% when highest and lowest intake groups were compared. The authors concluded that randomized controlled trials (RCTs) are needed. A large study comparing placebo to an all rac-alpha-tocopherol group consuming 400 IU/day reported no difference in bladder cancer cases. An inverse relationship between dietary vitamin E and lung cancer was reported in observational studies. The relative risk reduction was 16% when highest and lowest intake groups were compared. The benefit was progressive as dietary intake increased from 2 mg/day to 16 mg/day. The authors noted that the findings need to be confirmed by prospective studies. One such large trial, which compared 50 mg alpha-tocopherol to placebo in male tobacco smokers, reported no impact on lung cancer. A trial, which tracked people who chose to consume a vitamin E dietary supplement, reported an increased risk of lung cancer for those consuming more than 215 mg/day.
For prostate cancer, there are conflicting results. A meta-analysis based on serum alpha-tocopherol content reported an inverse correlation, with the difference between lowest and highest a 21% reduction in relative risk. In contrast, a meta-analysis of observational studies reported no relationship for dietary vitamin E intake. There were also conflicting results from large RCTs. The ATBC trial administered placebo or 50 mg/day alpha-tocopherol to male tobacco smokers for 5 to 8 years and reported a 32% decrease in the incidence of prostate cancer. Conversely, the SELECT trial of selenium and vitamin E for prostate cancer enrolled men ages 55 or older, mostly non-smokers, to consume a placebo or a 400 IU/day dietary supplement. It reported relative risk as a statistically significant 17% higher for the vitamin group.
For colorectal cancer, a systematic review identified RCTs of vitamin E and placebo followed for 7–10 years. There was a non-significant 11% decrease in relative risk. The SELECT trial (men over 55 years, placebo or 400 IU/day) also reported on colorectal cancer. There was a non-significant 3% increase in adenoma occurrence compared to placebo. The Women's Health Study compared placebo to 600 IU of natural-source vitamin E on alternate days for an average of 10.1 years. There were no significant differences for incidences of all types of cancer, cancer deaths, or for breast, lung or colon cancers.
Potential confounding factors are the form of vitamin E used in prospective studies and the amounts. Synthetic, racemic mixtures of vitamin E isomers are not bioequivalent to natural, non-racemic mixtures, yet are widely used in clinical trials and as dietary supplement ingredients. One review reported a modest increase in cancer risk with vitamin E supplementation while stating that more than 90% of the cited clinical trials used the synthetic, racemic form dl-alpha-tocopherol.
Cancer health claims
The U.S. Food and Drug Administration initiated a process of reviewing and approving food and dietary supplement health claims in 1993. Reviews of petitions results in proposed claims being rejected or approved. If approved, specific wording is allowed on package labels. In 1999, a second process for claims review was created. If there is not a scientific consensus on the totality of the evidence, a Qualified Health Claim (QHC) may be established. The FDA does not "approve" qualified health claim petitions. Instead, it issues a Letter of Enforcement Discretion that includes very specific claim language and the restrictions on using that wording. The first QHCs relevant to vitamin E were issued in 2003: "Some scientific evidence suggests that consumption of antioxidant vitamins may reduce the risk of certain forms of cancer." In 2009, the claims became more specific, allowing that vitamin E might reduce the risk of renal, bladder and colorectal cancers, but with required mention that the evidence was deemed weak and the claimed benefits highly unlikely. A petition to add brain, cervical, gastric and lung cancers was rejected. A further revision, May 2012, allowed that vitamin E may reduce risk of renal, bladder and colorectal cancers, with a more concise qualifier sentence added: "FDA has concluded that there is very little scientific evidence for this claim." Any company product label making the cancer claims has to include a qualifier sentence. The European Food Safety Authority (EFSA) reviews proposed health claims for the European Union countries. As of March 2018, EFSA has not evaluated any vitamin E and cancer prevention claims.
Cataracts
A meta-analysis from 2015 reported that for studies which reported serum tocopherol, higher serum concentration was associated with a 23% reduction in relative risk of age-related cataracts (ARC), with the effect due to differences in nuclear cataract rather than cortical or posterior subcapsular cataract – the three major classifications of age-related cataracts. However, this article and a second meta-analysis reporting on clinical trials of alpha-tocopherol supplementation reported no statistically significant change to risk of ARC when compared to placebo.
Cardiovascular diseases
Research on the effects of vitamin E on cardiovascular disease has produced conflicting results. In theory, oxidative modification of LDL-cholesterol promotes blockages in coronary arteries that lead to atherosclerosis and heart attacks, so vitamin E functioning as an antioxidant would reduce oxidized cholesterol and lower risk of cardiovascular disease. Vitamin E status has also been implicated in the maintenance of normal endothelial cell function of cells lining the inner surface of arteries, anti-inflammatory activity and inhibition of platelet adhesion and aggregation. An inverse relation has been observed between coronary heart disease and the consumption of foods high in vitamin E, and also higher serum concentration of alpha-tocopherol. In one of the largest observational studies, almost 90,000 healthy nurses were tracked for eight years. Compared to those in the lowest fifth for reported vitamin E consumption (from food and dietary supplements), those in the highest fifth were at a 34% lower risk of major coronary disease. The problem with observational studies is that these cannot confirm a relation between the lower risk of coronary heart disease and vitamin E consumption because of confounding factors. Diet higher in vitamin E may also be higher in other, unidentified components that promote heart health, or people choosing such diets may be making other healthy lifestyle choices.
There is some supporting evidence from randomized clinical trials (RCTs). A meta-analysis on the effects of alpha-tocopherol supplementation in RCTs on aspects of cardiovascular health reported that when consumed without any other antioxidant nutrient, the relative risk of heart attack was reduced by 18%. The results were not consistent for all of the individual trials incorporated into the meta-analysis. For example, the Physicians' Health Study II did not show any benefit after 400 IU every other day for eight years, for heart attack, stroke, coronary mortality or all-cause mortality. The HOPE/HOPE-TOO trial, which enrolled people with pre-existing vascular disease or diabetes into a multi-year trial of 400 IU/day, reported a higher risk of heart failure in the alpha-tocopherol group.
The effects of vitamin E supplementation on incidence of stroke were summarized in 2011. There were no significant benefits for vitamin E versus placebo. Subset analysis for ischaemic stroke, haemorrhagic stroke, fatal stroke, non-fatal stroke – all no significant difference in risk. Likewise for subset analysis of natural or synthetic vitamin E, or only above or below 300 IU/day, or whether the enrolled people were healthy or considered to be at higher than normal risk. The authors concluded that there was a lack of clinically important benefit of vitamin E supplementation in the prevention of stroke. One large, multi-year study in which post-menopausal women consumed either placebo or 600 IU of natural-sourced vitamin E on alternate days reported no effect on stroke, but did report a 21% reduction in relative risk of developing a deep vein clot or pulmonary embolism. The beneficial effect was strongest is the subset of women who had a history of a prior thrombotic event or who were genetically coded for clot risk (factor V Leiden or prothrombin mutation).
Cardiovascular health claims
In 2001, the U.S. Food and Drug Administration rejected proposed health claims for vitamin E and cardiovascular health. The U.S. National Institutes of Health reviewed literature published up to 2008 and concluded "In general, clinical trials have not provided evidence that routine use of vitamin E supplements prevents cardiovascular disease or reduces its morbidity and mortality." The European Food Safety Authority (EFSA) reviews proposed health claims for the European Union countries. In 2010, the EFSA reviewed and rejected claims that a cause and effect relationship has been established between the dietary intake of vitamin E and maintenance of normal cardiac function or of normal blood circulation.
Nonalcoholic fatty liver disease
A meta-analysis reported that vitamin E significantly reduced elevated liver enzymes, steatosis, inflammation and fibrosis, suggesting that the vitamin may be useful for treatment of nonalcoholic fatty liver disease (NAFLD) and the more extreme subset known as nonalcoholic steatohepatitis (NASH).
Parkinson's disease
There is an observed inverse correlation seen with dietary vitamin E, but no confirming evidence from placebo-controlled clinical trials. A meta-analysis published in 2005 concluded that diets higher in vitamin E content lowered risk of developing Parkinson's disease. From what appears to be the only clinical trial of tocopherol supplementation in people with early Parkinson's disease, 2,000 IU/day for 14 months had no effect on rate of disease progression.
Pregnancy
Antioxidant vitamins as dietary supplements have been proposed as having benefits if consumed during pregnancy. For the combination of vitamin E with vitamin C supplemented to pregnant women, a Cochrane review concluded that the data do not support vitamin E supplementation – majority of trials alpha-tocopherol at 400 IU/day plus vitamin C at 1,000 mg/day – as being efficacious for reducing risk of stillbirth, neonatal death, preterm birth, preeclampsia or any other maternal or infant outcomes, either in healthy women or those considered at risk for pregnancy complications. The review identified only three small trials in which vitamin E was supplemented without co-supplementation with vitamin C. None of these trials reported any clinically meaningful information.
Topical
Although there is widespread use of tocopheryl acetate as a topical medication, with claims for improved wound healing and reduced scar tissue, reviews have repeatedly concluded that there is insufficient evidence to support these claims. There are reports of vitamin E-induced allergic contact dermatitis from use of vitamin-E derivatives such as tocopheryl linoleate and tocopherol acetate in skin care products. Incidence is low despite widespread use.
Vaping-associated lung injury
On 5 September 2019, the US Food and Drug Administration announced that 10 out of 18, or 56% of the samples of vape liquids sent in by states, linked to recent vaping related lung disease outbreak in the United States, tested positive for vitamin E acetate which had been used as a thickening agent by illicit THC vape cartridge manufacturers. On 8 November 2019, the Centers for Disease Control and Prevention identified vitamin E acetate as a very strong culprit of concern in the vaping-related illnesses, but has not ruled out other chemicals or toxicants as possible causes. These findings were based on fluid samples from the lungs of 29 patients with vaping-associated pulmonary injury, which provided direct evidence of vitamin E acetate at the primary site of injury in all the 29 lung fluid samples tested.
History
Vitamin E was discovered in 1922 by Herbert McLean Evans and Katharine Scott Bishop and first isolated in a pure form by Evans and Gladys Anderson Emerson in 1935 at the University of California, Berkeley. Because the vitamin activity was first identified as a dietary fertility factor (in rats) it was given the name "tocopherol" from the Greek words "τόκος" [tókos, birth], and "φέρειν", [phérein, to bear or carry] meaning in sum "to carry a pregnancy," with the ending "-ol" signifying its status as a chemical alcohol. George M. Calhoun, Professor of Greek at the University of California, was credited with helping with the naming process. Erhard Fernholz elucidated its structure in 1938 and shortly afterwards the same year, Paul Karrer and his team first synthesized it.
Nearly 50 years after the discovery of vitamin E an editorial in the Journal of the American Medical Association titled "Vitamin in search of a disease" read in part "...research revealed many of the vitamin's secrets, but no certain therapeutic use and no definite deficiency disease in man." The animal discovery experiments had been a requirement for successful pregnancy, but no benefits were observed for women prone to miscarriage. Evidence for vascular health was characterized as unconvincing. The editorial closed with mention of some preliminary human evidence for protection against hemolytic anemia in young children.
A role for vitamin E in coronary heart disease was first proposed in 1946 by Evan Shute and colleagues. More cardiovascular work from the same research group followed, including a proposal that megadoses of vitamin E could slow down and even reverse the development of atherosclerosis. However, a 2004 meta-analysis showed no association between vitamin E supplementation and cardiovascular events (nonfatal stroke or myocardial infarction) or cardiovascular mortality. There is a long history of belief that topical application of vitamin E containing oil benefits burn and wound healing. This belief persists even though scientific reviews repeatedly refuted this claim.
The role of vitamin E in infant nutrition has a long research history. From 1949 onward there were trials with premature infants suggesting that oral alpha-tocopherol was protective against edema, intracranial hemorrhage, hemolytic anemia and retrolental fibroplasia. A 2003 Cochrane review concluded that vitamin E supplementation in preterm infants reduced the risk of intercranial hemorrhage and retinopathy, but noted an increased risk of sepsis.
See also
Tocopherol
Tocotrienol
References
External links
Bicyclic compounds
Food antioxidants |
null | null | Root beer | eng_Latn | Root beer is a sweet North American soft drink traditionally made using the root bark of the sassafras tree Sassafras albidum or the vine of Smilax ornata (known as sarsaparilla, also used to make a soft drink, sarsaparilla) as the primary flavor. Root beer is typically but not exclusively non-alcoholic, caffeine-free, sweet, and carbonated. Like beer, it usually has a thick and foamy head. A well-known use is to add vanilla ice cream to make a root beer float.
Since safrole, a key component of sassafras, was banned by the U.S. Food and Drug Administration in 1960 due to its carcinogenicity, most commercial root beers have been flavored using artificial sassafras flavoring, but a few (e.g. Hansen's) use a safrole-free sassafras extract.
Major root beer producers include A&W, Barq's, Dad's, Hires, and Mug.
History
Sassafras root beverages were made by indigenous peoples of the Americas for culinary and medicinal reasons before the arrival of Europeans in North America. European culinary techniques have been applied to making traditional sassafras-based beverages similar to root beer since the 16th century.
Root beer has been sold in confectionery stores since the 1840s, and written recipes for root beer have been documented since the 1860s. It possibly was combined with soda as early as the 1850s, and root beer sold in stores was most often sold as a syrup rather than a ready-made beverage.
The tradition of brewing root beer is thought to have evolved out of other small beer traditions that produced fermented drinks with very low alcohol content that were thought to be healthier to drink than possibly tainted local sources of drinking water, and enhanced by the medicinal and nutritional qualities of the ingredients used.
Beyond its aromatic qualities, the medicinal benefits of sassafras were well known to both Native Americans and Europeans, and druggists began marketing root beer for its medicinal qualities.
Pharmacist Charles Elmer Hires was the first to successfully market a commercial brand of root beer. Hires developed his root tea made from sassafras in 1875, debuted a commercial version of root beer at the Philadelphia Centennial Exposition in 1876, and began selling his extract. Hires was a teetotaler who wanted to call the beverage "root tea". However, his desire to market the product to Pennsylvania coal miners caused him to call his product "root beer", instead.
In 1886, Hires began to bottle a beverage made from his famous extract. By 1893, root beer was distributed widely across the United States. Non-alcoholic versions of root beer became commercially successful, especially during Prohibition.
Not all traditional or commercial root beers were sassafras-based. One of Hires's early competitors was Barq's, which began selling its sarsaparilla-based root beer in 1898 and was labeled simply as "Barq's".
In 1919, Roy Allen opened his root-beer stand in Lodi, California, which led to the development of A&W Root Beer. One of Allen's innovations was that he served his homemade root beer in cold, frosty mugs. IBC Root Beer is another brand of commercially-produced root beer that emerged during this period and is still well-known today.
Safrole, the aromatic oil found in sassafras roots and bark that gave traditional root beer its distinctive flavor, was banned for commercially mass-produced foods and drugs by the FDA in 1960. Laboratory animals that were given oral doses of sassafras tea or sassafras oil that contained large doses of safrole developed permanent liver damage or various types of cancer. While sassafras is no longer used in commercially produced root beer and is sometimes replaced with artificial flavors, natural extracts with the safrole distilled and removed are available.
Traditional method
One traditional recipe for making root beer involves cooking a syrup from molasses and water, letting the syrup cool for three hours, and combining it with the root ingredients (including sassafras root, sassafras bark, and wintergreen). Yeast was added, and the beverage was left to ferment for 12 hours, after which it was strained and rebottled for secondary fermentation. This recipe usually resulted in a beverage of 2% alcohol or less, although the recipe could be modified to produce a more alcoholic beverage.
Foam
Root beer is identified by its classic foam, appearing as white bubbles on the surface of the drink. Root beer was originally made partially with sassafras root bark (and sarsaparilla, etc) which naturally foamed, giving it its distinctive look. Root beer manufacturers initially carbonated the drink to add bubbles, later adding a surfactant to lower the surface tension and let the bubbles last longer. Different brands of root beer have slightly different foams, giving each a different identity.
Ingredients
Commercial root beer is now produced in Canada and every U.S. state. Although this beverage's popularity is greatest in North America, brands are produced and imported in other countries, including Australia, the United Kingdom, Malaysia, Argentina, Germany, the Philippines, Singapore, Taiwan, South Korea, Indonesia, Sweden, Vietnam, and Thailand. The flavor of these beverages may vary from typical North American versions, or be similar to those found in North America. While no standard recipe exists, the primary ingredients in modern root beer are filtered water, sugar, and safrole-free sassafras extract, which complements other flavors. Common flavorings are vanilla, caramel, wintergreen, black cherry bark, licorice root, sarsaparilla root, nutmeg, acacia, anise, molasses, cinnamon, sweet birch, and honey. Soybean protein, or yucca is sometimes used to create a foamy quality, and caramel coloring is used to make the beverage brown.
Ingredients in early and traditional root beers include allspice, birch bark, coriander, juniper, ginger, wintergreen, hops, burdock root, dandelion root, spikenard, pipsissewa, guaiacum chips, sarsaparilla, spicewood, wild cherry bark, yellow dock, prickly ash bark, sassafras root, vanilla beans, dog grass, molasses and licorice. Many of these ingredients are still used in traditional and commercially produced root beer today, which is often thickened, foamed or carbonated.
Most major brands other than Barq's are caffeine-free (Barq's contains about 1.8 mg of caffeine per fluid ounce).
Root beer can be made at home with processed extract obtained from a factory, or it can also be made from herbs and roots that have not yet been processed. Alcoholic and non-alcoholic traditional root beers make a thick and foamy head when poured, often enhanced by the addition of yuca extract, soybean protein, or other thickeners.
Alcoholic root beers produced in the 2000s have included Small Town Brewery's Not Your Father's Root Beer; Coney Island Brewing Co.'s hard root beer; and Best Damn Brewing Co.'s Best Damn Root Beer.
Common ingredients
Roots and herbs
Spices
Other ingredients
See also
Apple Beer
Beer
Beverage
Birch beer
:Category:Root beer stands
Cream soda
Dandelion and burdock
Ginger beer
Horehound beer
Julmust
Lewis and Clark
List of brand name soft drinks products
List of soft drink flavors
Malta (soft drink)
Malzbier
Moxie
Root beer float
Sarsaparilla (soft drink) – a similar, although distinct, beverage
Spruce beer
References
External links
American drinks
Non-timber forest products
Cuisine of Philadelphia |
null | null | Fast food restaurant | eng_Latn | A fast food restaurant, also known as a quick service restaurant (QSR) within the industry, is a specific type of restaurant that serves fast food cuisine and has minimal table service. The food served in fast food restaurants is typically part of a "meat-sweet diet", offered from a limited menu, cooked in bulk in advance and kept hot, finished and packaged to order, and usually available for take away, though seating may be provided. Fast food restaurants are typically part of a restaurant chain or franchise operation that provides standardized ingredients and/or partially prepared foods and supplies to each restaurant through controlled supply channels. The term "fast food" was recognized in a dictionary by Merriam–Webster in 1951.
Arguably, the first fast food restaurants originated in the United States with White Castle in 1921. Today, American-founded fast food chains such as McDonald's (est. 1940) and KFC (est. 1952) are multinational corporations with outlets across the globe.
Variations on the fast food restaurant concept include fast casual restaurants and catering trucks. Fast casual restaurants have higher sit-in ratios, offering a hybrid between counter-service typical at fast food restaurants and a traditional table service restaurant. Catering trucks (also called food trucks) often park just outside worksites and are popular with factory workers.
History
In 1896, the first self-service restaurant (the "Stollwerck-Automatenrestaurant") opened in Berlin's Leipziger Straße.
United States
Some trace the modern history of fast food in the United States to 7 July 1912, with the opening of a fast food restaurant called the Automat in New York. The Automat was a cafeteria with its prepared foods behind small glass windows and coin-operated slots. Joseph Horn and Frank Hardart had already opened the first Horn & Hardart Automat in Philadelphia in 1902, but their "Automat" at Broadway and 13th Street, in New York City, created a sensation. Numerous Automat restaurants were built around the country to deal with the demand. Automats remained extremely popular throughout the 1920s and 1930s. The company also popularized the notion of "take-out" food, with their slogan "Less work for Mother".
Most historians agree that the American company White Castle was the first fast food outlet, starting in Wichita, Kansas in 1916 with food stands and founding in 1921, selling hamburgers for five cents apiece from its inception and spawning numerous competitors and emulators. What is certain, however, is that White Castle made the first significant effort to standardize the food production in, look of, and operation of fast food hamburger restaurants. William Ingram's and Walter Anderson's White Castle System created the first fast food supply chain to provide meat, buns, paper goods, and other supplies to their restaurants, pioneered the concept of the multi-state hamburger restaurant chain, standardized the look and construction of the restaurants themselves, and even developed a construction division that manufactured and built the chain's prefabricated restaurant buildings. The McDonald's Speedee Service System and, much later, Ray Kroc's McDonald's outlets and Hamburger University all built on principles, systems and practices that White Castle had already established between 1923 and 1932.
The hamburger restaurant most associated by the public with the term "fast food" was created by two brothers originally from Nashua, New Hampshire. Richard and Maurice McDonald opened a barbecue drive-in in 1940 in the city of San Bernardino, California. After discovering that most of their profits came from hamburgers, the brothers closed their restaurant for three months and reopened it in 1948 as a walk-up stand offering a simple menu of hamburgers, french fries, shakes, coffee, and Coca-Cola, served in disposable paper wrapping. As a result, they could produce hamburgers and fries constantly, without waiting for customer orders, and could serve them immediately; hamburgers cost 15 cents, about half the price at a typical diner. Their streamlined production method, which they named the "Speedee Service System" was influenced by the production line innovations of Henry Ford.
By 1954, The McDonald brothers' stand was restaurant equipment manufacturer Prince Castle's biggest purchaser of milkshake blending machines. Prince Castle salesman Ray Kroc traveled to California to discover why the company had purchased almost a dozen of the units as opposed to the normal one or two found in most restaurants of the time. Enticed by the success of the McDonald's concept, Kroc signed a franchise agreement with the brothers and began opening McDonald's restaurants in Illinois. By 1961, Kroc had bought out the brothers and created what is now the modern McDonald's Corporation. One of the major parts of his business plan was to promote cleanliness of his restaurants to growing groups of Americans that had become aware of food safety issues. As part of his commitment to cleanliness, Kroc often took part in cleaning his own Des Plaines, Illinois outlet by hosing down the garbage cans and scraping gum off the cement. Another concept Kroc added was great swaths of glass which enabled the customer to view the food preparation, a practice still found in chains such as Krispy Kreme. A clean atmosphere was only part of Kroc's grander plan which separated McDonald's from the rest of the competition and attributes to their great success. Kroc envisioned making his restaurants appeal to suburban families.
At roughly the same time as Kroc was conceiving what eventually became McDonald's Corporation, two Miami, Florida businessmen, James McLamore and David Edgerton, opened a franchise of the predecessor to what is now the international fast food restaurant chain Burger King. McLamore had visited the original McDonald's hamburger stand belonging to the McDonald brothers; sensing potential in their innovative assembly line-based production system, he decided he wanted to open a similar operation of his own. The two partners eventually decided to invest their money in Jacksonville, Florida-based Insta-Burger King. Originally opened in 1953, the founders and owners of the chain, Kieth G. Kramer and his wife's uncle Matthew Burns, opened their first stores around a piece of equipment known as the Insta-Broiler. The Insta-Broiler oven proved so successful at cooking burgers, they required all of their franchises to carry the device. By 1959 McLamore and Edgarton were operating several locations within the Miami-Dade area and were growing at a fast clip. Despite the success of their operation, the partners discovered that the design of the insta-broiler made the unit's heating elements prone to degradation from the drippings of the beef patties. The pair eventually created a mechanized gas grill that avoided the problems by changing the way the meat patties were cooked in the unit. After the original company began to falter in 1959, it was purchased by McLamore and Edgerton who renamed the company Burger King.
While fast food restaurants usually have a seating area in which customers can eat the food on the premises, orders are designed to be taken away, and traditional table service is rare. Orders are generally taken and paid for at a wide counter, with the customer waiting by the counter for a tray or container for their food. A "drive-through" service can allow customers to order and pick up food from their cars.
Nearly from its inception, fast food has been designed to be eaten "on the go" and often does not require traditional cutlery and is eaten as a finger food. Common menu items at fast food outlets include fish and chips, sandwiches, pitas, hamburgers, fried chicken, french fries, chicken nuggets, tacos, pizza, and ice cream, although many fast food restaurants offer "slower" foods like chili, mashed potatoes, and salads.
Cuisine
Modern commercial fast food is highly processed and prepared on a large scale from bulk ingredients using standardized cooking and production methods and equipment. It is usually rapidly served in cartons, bags, or in a plastic wrapping, in a fashion which reduces operating costs by allowing rapid product identification and counting, promoting longer holding time, avoiding transfer of bacteria, and facilitating order fulfillment. In most fast food operations, menu items are generally made from processed ingredients prepared at central supply facilities and then shipped to individual outlets where they are cooked (usually by grill, microwave, or deep-frying) or assembled in a short amount of time either in anticipation of upcoming orders (i.e., "to stock") or in response to actual orders (i.e., "to order"). Following standard operating procedures, pre-cooked products are monitored for freshness and disposed of if holding times become excessive. This process ensures a consistent level of product quality, and is key to delivering the order quickly to the customer and avoiding labor and equipment costs in the individual stores.
Because of commercial emphasis on taste, speed, product safety, uniformity, and low cost, fast food products are made with ingredients formulated to achieve an identifiable flavor, aroma, texture, and "mouth feel" and to preserve freshness and control handling costs during preparation and order fulfillment. This requires a high degree of food engineering. The use of additives, including salt, sugar, flavorings and preservatives, and processing techniques may limit the nutritional value of the final product.
Value meals
A value meal is a group of menu items offered together at a lower price than they would cost individually. A hamburger, side of fries, and drink commonly constitute a value meal—or combo depending on the chain. Value meals at fast food restaurants are common as a merchandising tactic to facilitate bundling, up-selling, and price discrimination. Most of the time they can be upgraded to a larger side and drink for a small fee. The perceived creation of a "discount" on individual menu items in exchange for the purchase of a "meal" is also consistent with the loyalty marketing school of thought.
Technology
To make quick service possible and to ensure accuracy and security, many fast food restaurants have incorporated hospitality point of sale systems. This makes it possible for kitchen crew people to view orders placed at the front counter or drive through in real time. Wireless systems allow orders placed at drive through speakers to be taken by cashiers and cooks. Drive through and walk through configurations will allow orders to be taken at one register and paid at another. Modern point of sale systems can operate on computer networks using a variety of software programs. Sales records can be generated and remote access to computer reports can be given to corporate offices, managers, troubleshooters, and other authorized personnel.
Food service chains partner with food equipment manufacturers to design highly specialized restaurant equipment, often incorporating heat sensors, timers, and other electronic controls into the design. Collaborative design techniques, such as rapid visualization and computer-aided design of restaurant kitchens are now being used to establish equipment specifications that are consistent with restaurant operating and merchandising requirements.
Business
Consumer spending
In the United States, consumers spent about US$110 billion on fast food in 2000 (which increased from $6 billion in 1970). The National Restaurant Association forecasts that fast food restaurants in the US will reach $142 billion in sales in 2006, a 5% increase over 2005. In comparison, the full-service restaurant segment of the food industry is expected to generate $173 billion in sales. Fast food has been losing market share to so-called fast casual restaurants, which offer more robust and expensive cuisines.
Major international brands
McDonald's, a fast food supplier, opened its first franchised restaurant in the US in 1955 (1974 in the UK). It has become a phenomenally successful enterprise in terms of financial growth, brand-name recognition, and worldwide expansion. Ray Kroc, who bought the franchising license from the McDonald brothers, pioneered concepts which emphasized standardization. He introduced uniform products, identical in all respects at each outlet, to increase sales. Kroc also insisted on cutting food costs as much as possible, eventually using the McDonald's Corporation's size to force suppliers to conform to this ethos.
Other prominent international fast food companies include Burger King, the number two hamburger chain in the world, known for promoting its customized menu offerings (Have it Your Way). Another international fast food chain is KFC, which sells chicken-related products and is the number 1 fast food company in the People's Republic of China.
Franchising
A fast food chain restaurant is generally owned either by the parent company of the fast food chain or a franchisee – an independent party given the right to use the company's trademark and trade name. In the latter case, a contract is made between the franchisee and the parent company, typically requiring the franchisee to pay an initial, fixed fee in addition to a continual percentage of monthly sales. Upon opening for business, the franchisee oversees the day-to-day operations of the restaurant and acts as a manager of the store. Once the contract expires, the parent company may choose to "renew the contract, sell the franchise to another franchisee, or operate the restaurant itself." In most fast food chains, the number of franchised locations exceeds the number of company owned locations.
Fast food chains rely on consistency and uniformity, in internal operations and brand image, across all of their restaurant locations in order to convey a sense of reliability to their customers. This sense of reliability coupled with a positive customer experience brings customers to place trust in the company. This sense of trust leads to increased customer loyalty which gives the company a source of recurring business. When a person is presented with a choice of different restaurants to eat at, it is much easier for them to stick with what they know, rather than to take a gamble and dive into the unknown.
Due to the importance of consistency, most companies set standards unifying their various restaurant locations with a set of common rules and regulations. Parent companies often rely on field representatives to ensure that the practices of franchised locations are consistent with the company's standards. However, the more locations a fast food chain has, the harder it is for the parent company to guarantee that these standards are being followed. Moreover, it is much more expensive to discharge a franchisee for noncompliance with company standards, than it is to discharge an employee for that same reason. As a consequence, parent companies tend to deal with franchisee violations in a more relaxed manner.
Many companies also adapt to their different local areas to support the needs of the customers. Sometimes it is necessary for a franchisee to modify the way the restaurant/store runs to meet the needs of local customers. As referenced in Bodey's "Localization and Customer Retention for Franchise Service Systems" article, J. L. Bradach claims that a franchise will either use the tactical or strategic local response. Tactical applies to accounting for hiring of personnel and suppliers as well as financial decisions to adapt to the local customers. Strategic applies to the specific characteristics of the franchise that will change from the basic format followed by all to fit in the local area.
For the most part, someone visiting a McDonald's in the United States will have the same experience as someone visiting a McDonald's in Japan. The interior design, the menu, the speed of service, and the taste of the food will all be very similar. However, some differences do exist to tailor to particular cultural differences. For example, in October 2005 during a midst of plummeting sales in Japan, McDonald's added a shrimp burger to the Japanese menu. The choice to introduce a shrimp burger was no coincidence, as a 1989 study stated that world consumption of shrimp was "led by Japan."
In March 2010, Taco Bell opened their first restaurant in India. Because non-consumption of beef is a cultural norm in light of India's Dharmic beliefs, Taco Bell had to tailor its menu to the dietary distinctions of Indian culture by replacing all of the beef with chicken. By the same token, completely meatless options were introduced to the menu due to the prevalence of vegetarianism throughout the country.
Countries
Multinational corporations typically modify their menus to cater to local tastes, and most overseas outlets are owned by native franchisees. McDonald's in India, for example, uses chicken and paneer rather than beef and pork in their burgers because Hinduism traditionally forbids eating beef. In Israel some McDonald's restaurants are kosher and respect the Jewish Shabbat; there is also a kosher McDonald's in Argentina. In Egypt, Indonesia, Morocco, Saudi Arabia, Malaysia, Pakistan, and Singapore, all menu items are halal.
North America
Many fast food operations have more local and regional roots, such as White Castle in the Midwest United States, along with Hardee's (owned by CKE Restaurants, which also owns Carl's Jr., whose locations are primarily on the United States West Coast); Krystal, Bojangles' Famous Chicken 'n Biscuits, Cook Out, and Zaxby's restaurants in the American Southeast; Raising Cane's in Louisiana and other mostly Southern states; Hot 'n Now in Michigan and Wisconsin; In-N-Out Burger (in California, Arizona, Nevada, Utah, and Texas, with a few locations in Oregon) and Original Tommy's chains in Southern California; Dick's Drive-In in Seattle, Washington and Arctic Circle in Utah and other western states; Halo Burger around Flint, Michigan and Burgerville in the Portland, Oregon area. Also, Whataburger is a popular burger chain in the American South, and Jack in the Box is located in the West and South. Canada pizza chains Topper's Pizza and Pizza Pizza are primarily located in Ontario. Coffee chain Country Style operates only in Ontario, and competes with the famous coffee and donut chain Tim Hortons. Maid-Rite restaurant is one of the oldest chain fast food restaurants in the United States. Founded in 1926, their specialty is a loose meat hamburger. Maid-Rites can be found in the midwest - mainly Iowa, Minnesota, Illinois, and Missouri.
International brands dominant in North America include McDonald's, Burger King and Wendy's, the number three burger chain in the USA; Dunkin' Donuts, a New England-based chain; automobile oriented Sonic Drive-In's from Oklahoma City; Starbucks, Seattle-born coffee-based fast food beverage corporation; KFC and Taco Bell, which are both part of the largest restaurant conglomerate in the world, Yum! Brands; and Domino's Pizza, a pizza chain known for popularizing home delivery of fast food.
Subway is known for their sub sandwiches and are the largest restaurant chain to serve such food items. Quiznos a Denver-based sub shop is another fast-growing sub chain, yet with over 6,000 locations it is still far behind Subway's 34,000 locations. Other smaller sub shops include Blimpie, Jersey Mike's Subs, Mr. Goodcents, Jimmy John's, and Firehouse.
A&W Restaurants was originally a United States and Canada fast food brand, but it is currently an International fast food corporation in several countries.
In Canada the majority of fast food chains are American owned or were originally American owned but have since set up a Canadian management/headquarters locations such as Panera Bread, Chipotle Mexican Grill, Five Guys, and Carl's Jr. Although the case is usually American fast food chains expanding into Canada, Canadian chains such as Tim Hortons have expanded into 22 states in the United States, but are more prominent in border states such as New York and Michigan. Tim Hortons has started to expand to other countries outside of North America. The Pita Pit franchise originated in Canada and has expanded to the United States and other Countries. The Canadian Extreme Pita franchise sells low fat and salt pita sandwiches with stores in the larger Canadian cities. Other Canadian fast food chains such as Manchu Wok serve North American style Asian foods; this company is located mainly in Canada and the US, with other outlets on US military bases on other continents. Harvey's is a Canadian-only burger restaurant chain, present in every province.
Oceania
Australia
Australia's fast food market began in 1968, with the opening of several American franchises including McDonald's and KFC. Pizza Hut was introduced in April 1970, and Burger King followed. However, the Burger King market found that this name was already a registered trademark to a takeaway food shop in Adelaide. Thus, the Burger King Australian market was forced to pick another name, selecting the Hungry Jack's brand name. Prior to this, the Australian fast food market consisted primarily of privately owned take-away shops.
New Zealand
In New Zealand, the fast food market began in the 1970s with KFC (opened 1971), Pizza Hut (1974), and McDonald's (1976), and all three remain popular today. Burger King and Domino's entered the market later in the 1990s. Australian pizza chains Eagle Boys and Pizza Haven also entered the market in the 1990s, but their New Zealand operations were later sold to Pizza Hut and Domino's.
A few fast food chains have been founded in New Zealand, including Burger Fuel (founded 1995), Georgie Pie (founded 1977, but closed 1998 after falling into financial trouble and being bought out by McDonald's) and Hell Pizza (founded 1996).
Europe
United Kingdom
The United Kingdom's signature type of fast food restaurant is a fish and chip shop, which specializes in fish and chips and also other foods such as kebabs and burgers. Fish and chip shops are usually owned independently. Many other types of home based fast food operations were closed in the 1970s and 1980s after McDonald's became the number one outlet in the market.
However, brands like Wimpy still remain, although the majority of branches became Burger King in 1989.
Netherlands
In the Netherlands, walk-up fast food restaurants with automatiek, a typical Dutch vending machine, are found throughout the country, most notably in Amsterdam. In this automatic format, a counter is available for purchasing French fries, beverages, krokets, frikandellen, kaassoufflés and hamburgers and other snacks can be bought from the automats. FEBO is one of the largest of such types of fast-food restaurants with automats.
Ireland
In addition to home-grown chains such as Supermac's, numerous American chains such as McDonald's and Burger King have also established a presence in Ireland. In 2015, a study developed by Treated.com was published in the Irish Times, which named Swords in County Dublin as Ireland's 'fast food capital'.
Asia
Japan
American chains such as Domino's Pizza, McDonald's, Pizza Hut, and KFC have a big presence in Japan, but local gyudon chains such as Sukiya, Matsuya and Yoshinoya also blanket the country. Japan has its own burger chains including MOS Burger, Lotteria and Freshness Burger.
Taiwan
Notable Taiwanese fast-food restaurants include 85C Bakery Cafe and TKK Fried Chicken.
India
The major fast food chains in India that serve American fast food are KFC, McDonald's, Starbucks, Burger King, Subway, Pizza Hut, and Dominos. Most of these have had to make a lot of changes to their standard menus to cater to Indian food habits and taste preferences. Some emerging Indian food chains include Wow! Momo, Haldiram's, Faaso's and Café Coffee Day.
Food habits vary widely across states within India. While typical Idli and Dosa is fast food in Southern India, in Maharashtra it is misal-pav, pav-bhaji, and poha. Further north in Punjab and Haryana, chole-bhature are very popular and in Bihar and Jharkhand litti-chokha is their staple fast food.
Pakistan
Fast food In Pakistan varies. There are many international chains serving fast food, including Nandos, Burger King, KFC, McDonald's, Domino's Pizza, Fatburger, Dunkin' Donuts, Subway, Pizza Hut, Hardees, Telepizza, Steak Escape and Gloria Jean's Coffees. In addition to the international chains, in local cuisine people in Pakistan like to have biryani, bun kebabs, Nihari, kebab rolls etc. as fast food.
Philippines
In the Philippines, fast-food is the same as in the US. However, the only difference is that they serve Filipino dishes and a few American products being served Filipino-style. Jollibee is the leading fast food chain in the country with 1,000 stores nationwide.
Russia
Most international fast food chains like Subway, McDonald's, Burger King etc. are represented in major Russian cities. There are also local chains like Teremok specializing in Russian cuisine or having elements of it added into their menu.
Saudi Arabia
Saudi Arabia has many international fast food chains including KFC, Burger King, McDonald's and many others. However, the most popular fast food restaurant of Saudi Arabia is Albaik. Saudis regard Albaik as better than KFC.
Hong Kong
In Hong Kong, although McDonald's and KFC are quite popular, three major local fast food chains provide Hong Kong style fast food, namely Café de Coral, Fairwood, and Maxim MX. Café de Coral alone serves more than 300,000 customers daily. Unlike western fast food chains, these restaurants offer four different menus at different times of the day, namely breakfast, lunch, afternoon tea, and dinner. Siu mei is offered throughout the day.
Dai pai dong and traditional Hong Kong street food may be considered close relatives of the conventional fast food outlet.
Israel
In Israel, local burger chain Burger Ranch is popular as are McDonald's and Burger King. Domino's Pizza is also a popular fast food restaurant. Chains like McDonald's offer kosher branches. Non-kosher foods such as cheeseburgers are rare in Israeli fast food chains, even in non-kosher branches. There are many small local fast food chains that serve pizza, hamburgers, sushi and local foods such as hummus, falafel and shawarma.
Africa
Nigeria
In Nigeria, Mr. Bigg's, Chicken Republic, Tantalizers, and Tastee Fried Chicken are the predominant fast food chains. KFC and Domino's Pizza have recently entered the country.
South Africa
KFC is the most popular fast food chain in South Africa according to a 2010 Sunday Times survey. Chicken Licken, Wimpy and Ocean Basket along with Nando's and Steers are examples of homegrown franchises that are highly popular within the country. McDonald's, Subway and Pizza Hut have a significant presence within South Africa.
Trends
Health concerns
Some of the large fast food chains are beginning to incorporate healthier alternatives in their menu, e.g., white meat, snack wraps, salads, and fresh fruit. However, some people see these moves as a tokenistic and commercial measure, rather than an appropriate reaction to ethical concerns about the world ecology and people's health. McDonald's announced that in March 2006, the chain would include nutritional information on the packaging of all of its products.
In September and October 2000, during the Starlink corn recalls, up to $50 million worth of corn-based foods were recalled from restaurants as well as supermarkets. The products contained Starlink genetically modified corn that was not approved for human consumption. It was the first-ever recall of a genetically modified food. The environmental group Friends of the Earth that had first detected the contaminated shells was critical of the FDA for not doing its own job.
Fast food is commonly blamed for the obesity epidemic in the United States today. 60% of Americans today are either overweight or obese. With obesity especially being seen among children, places like McDonald's and other fast food restaurants take the majority of the blame. 34% of children and adolescents consume fast food on any given day, while 80% of children claim that McDonald's is their favorite place to eat at. The number of children and adolescents as well as adults eating out every day is only seen to progress and rise. Research concludes that children and adolescents ranging from twelve to nineteen years old consume twice as many calories from fast food restaurants than children ranging from two to eleven years old.
The FDA found that trans fats raises the amount of cholesterol in blood, which raises the chance of developing heart disease which is known as one of the leading causes of death in the U.S. In a recent study, it was found that 11 out of 25 restaurants failed after tests on the use of antibiotics as well. Antibiotic-resistant infections affect at least 2 million Americans each year, which will leave at least 23,000 of those people to die.
Consumer appeal
Fast food outlets have become popular with consumers for several reasons. One is that through economies of scale in purchasing and producing food, these companies can deliver food to consumers at a very low cost. In addition, although some people dislike fast food for its predictability, it can be reassuring to a hungry person in a hurry or far from home.
In the post-World War II period in the United States, fast food chains like McDonald's rapidly gained a reputation for their cleanliness, fast service, and a child-friendly atmosphere where families on the road could grab a quick meal. Prior to the rise of the fast food chain restaurant, people generally had a choice between diners where the quality of the food was often questionable and service lacking, or high-end restaurants that were expensive and impractical for families with children. The modern, stream-lined convenience of the fast food restaurant provided a new alternative and appealed to Americans' instinct for ideas and products associated with progress, technology, and innovation. Fast food restaurants rapidly became the eatery "everyone could agree on", with many featuring child-size menu combos, play areas, and whimsical branding campaigns, like the iconic Ronald McDonald, designed to appeal to younger customers. Parents could have a few minutes of peace while children played or amused themselves with the toys included in their Happy Meal. There is a long history of fast food advertising campaigns, many of which are directed at children.
Fast food marketing largely focuses on children and teenagers. Popular methods of advertising include television, product placement in toys, games, educational materials, songs, and movies, character licensing and celebrity endorsements, and websites. Advertisements targeting children mainly focus on free toys, movie tie-ins and other giveaways. Fast food restaurants use kid's meals with toys, kid friendly mascots, vibrant colors, and play areas to draw children toward their products. . Children's power over their parents' purchases is estimated to total $300 to $500 billion every year. Fast food has become a part of American culture as a reward for children. To deny a child "desirable things" such as the advertised fast food restaurant can cause stigmatization of parents as the "mean parent" when it is common among other parents to comply with their child's desires.
The major focus on children by the fast food industry has created controversy due to the rising issue of child obesity in America. As a result of this focus, in 2008 a coalition was created and run by the Council of Better Business Bureaus called Children's Food and Beverage Advertising Initiative(CFBAI), to stop ads aimed at children or to promote only what the council dubs "better-for- you" products in ads directed towards children. However, it was not until 2011 that Congress requested guidelines be put in place by the CFBAI, FDA, Agriculture Department, and Centers for Disease Control. There are two basic requirements identified in the guidelines for foods that are advertised for children: (1) The food has to include healthful ingredients; (2) The food can't contain unhealthful amounts of sugar, Saturated fat, Trans fat, and salt. The guidelines are voluntary but companies experience heavy pressure to comply. Once a company complies they have 5–10 years to comply with the guidelines. Many fast food industries have started to comply with the guidelines. Although many companies have ways to go. In 2012 the fast food industry spent $4.6 billion to advertise unhealthy products to children and teens according to a report by the Yale Rudd Center for Food Policy & Obesity. There are points of progress that include healthier sides and beverages in most fast food restaurant kids' meals. The guidelines are interested in a healthier lifestyle for children and the growing problem of American obesity.
While it may not be so obvious today, fast food is on its way to healthier foods and menus. Some businesses have taken off and had success in this market with healthy foods and due to many health concerns and fast food stereotypes, it is emerging as an expected thing. According to Jason Daley, healthful fast food takeoff is explained by the fact that it has "ceased to be a thing-now, it's simply expected."
In other parts of the world, American and American-style fast food outlets have been popular for their quality, customer service, and novelty, even though they are often the targets of popular anger towards American foreign policy or globalization more generally. Many consumers nonetheless see them as symbols of the wealth, progress, and well-ordered openness of Western society and they therefore become trendy attractions in many cities around the world, particularly among younger people with more varied tastes.
Impact of fast food restaurant availability
Over time, fast food restaurants have been growing rapidly, especially in urban neighborhoods. According to US research, low-income and predominantly African-American neighborhoods have greater exposure to fast food outlets than higher income and predominantly white areas. This has put into question whether urbanized neighborhoods were targeted, which causes a more unhealthy group of people compared to people from a higher socioeconomic status. It has also been shown that there is a lower chance of finding a fast food restaurant in a suburban neighborhood. In a study of selected US locations, Morland et al. (2002) found the number of fast food restaurants and bars was inversely proportional to the wealth of the neighborhood, and that predominantly African-American residential areas were four times less likely to have a supermarket near them than predominantly white areas.
Innovations timeline
1872: Walter Scott of Providence, RI outfitted a horse-drawn lunch wagon with a simple kitchen, bringing hot dinners to workers
1916: Walter Anderson built the first White Castle in Wichita, KS in 1916, introducing the limited menu, high volume, low cost, high speed hamburger restaurant
1919: A&W Root Beer took its product out of the soda fountain and into a roadside stand
1921: A&W Root Beer began franchising its syrup
1930s: Howard Johnson's pioneered the concept of franchising restaurants, formally standardizing menus, signage, and advertising
1967: McDonald's opens its first restaurants outside the US.
Halal
The introduction of the halal option by some fast food companies saw the expansion of fast food chains into Muslim majority countries has resulted in a rise of restaurant options in non-western nations and has also increased revenue for some western restaurant chains. Some outlets offering Halal options include KFC, Nando's, Pizza Express, and Subway. McDonald's carried out a trial but decided that the cost of operations would be too high. There have also been court cases involving start-up businesses during attempts to alter the halal-certified method by machine killing, which is against the beliefs of some Muslims. However, the trend towards halal has been unpopular in some communities which have at times resulted in internet petitions.
Criticisms
The fast food industry is a popular target for critics, from anti-globalization activists like José Bové to vegetarian activist groups such as PETA as well as the workers themselves. A number of fast food worker strikes occurred in the United States in the 2010s.
In his best-selling 2001 book Fast Food Nation, investigative journalist Eric Schlosser leveled a broad, socioeconomic critique against the fast food industry, documenting how fast food rose from small, family-run businesses (like the McDonald brothers' burger joint) into large, multinational corporate juggernauts whose economies of scale radically transformed agriculture, meat processing, and labor markets in the late twentieth century. Schlosser argues that while the innovations of the fast food industry gave Americans more and cheaper dining options, it has come at the price of destroying the environment, economy, and small-town communities of rural America while shielding consumers from the real costs of their convenient meal, both in terms of health and the broader impact of large-scale food production and processing on workers, animals, and land.
The fast food industry is popular in the United States, the source of most of its innovation, and many major international chains are based there. Seen as symbols of US dominance and perceived cultural imperialism, American fast food franchises have often been the target of Anti-globalization protests and demonstrations against the US government. In 2005, for example, rioters in Karachi, Pakistan, who were initially angered because of the bombing of a Shiite mosque, destroyed a KFC restaurant.
Legal issues
In August 2002, a group of overweight children in New York City filed a class-action lawsuit against McDonald's Corporation seeking compensation for obesity-related health problems, improved nutritional labeling of McDonald's products, and funding for a program to educate consumers about the dangers of fast food. This provoked an intense, mostly negative response in the media with columnists calling this case a "cartoon of a lawsuit". Also suggesting the lawyers were poisoned to "get fat" off McDonald's.
This kind of litigation raises the important question of who, if anyone, ought to be held accountable for the economic and public health consequences of obesity.
In 2003, McDonald's was sued in a New York court by a family who claimed that the restaurant chain was responsible for their teenage daughter's obesity and attendant health problems. By manipulating food's taste, sugar and fat content, and directing their advertising to children, the suit argued that the company purposely misleads the public about the nutritional value of its product. A judge dismissed the case, but the fast food industry disliked the publicity of its practices, particularly the way it targets children in its advertising. Although further lawsuits have not materialized, the issue is kept alive in the media and political circles by those promoting the need for tort reform.
In response to this, the "Cheeseburger Bill" was passed by the US House of Representatives in 2004; it later stalled in the US Senate. The law was reintroduced in 2005, only to meet the same fate. This law was claimed to "[ban] frivolous lawsuits against producers and sellers of food and non-alcoholic drinks arising from obesity claims."
The bill arose because of an increase in lawsuits against fast food chains by people who claimed that eating their products made them obese, disassociating themselves from any of the blame.
See also
Fast food advertising
HACCP
List of fast food restaurant chains
List of hamburger restaurants
List of the largest fast food restaurant chains
Roadhouse (facility)
Sanitation Standard Operating Procedures
References
Further reading
Hogan, David. Selling 'em by the Sack: White Castle and the Creation of American Food. New York: New York University Press, 1997.
Kroc, Ray and Anderson, Robert Grinding It Out: The Making of McDonald's. Chicago: Contemporary Books, 1977.
Levinstein, Harvey. Paradox of Plenty: a Social History of Eating in Modern America. Berkeley: University of California P, 2003. 228–229.
Luxenberg, Stan. Roadside Empires: How the Chains Franchised America. New York: Viking, 1985.
Mcginley, Lou Ellen with Stephanie Spurr. "Honk for Service: A Man, A Tray and the Glory Days of the Drive-In Restaurant". Tray Days Publishing, 2004
Schlosser, Eric. "Fast Food Nation: The Dark Side of the All American Meal" HarperCollins Publishers, 2005
Schultz, Howard and Yang, Dori Jones. "Pour Your Heart Into It: How Starbucks Built a Company One Cup at a Time". Hyperion, 1999.
Restaurants by type
Restaurant |
null | null | Slave rebellion | eng_Latn | A slave rebellion is an armed uprising by enslaved people, as a way of fighting for their freedom. Rebellions of enslaved people have occurred in nearly all societies that practice slavery or have practiced slavery in the past. A desire for freedom and the dream of successful rebellion is often the greatest object of song, art, and culture amongst the enslaved population. Many of the events, however, are often violently opposed and suppressed by slaveholders.
The most successful slave rebellion in history was the 18th-century Haitian Revolution, led by Toussaint Louverture and later Jean-Jacques Dessalines who won the war against their French colonial rulers, which established the modern independent state of Haiti from the former French colony of Saint-Domingue. Another famous historic slave rebellion was led by the Roman slave Spartacus (c. 73–71 BC). In the ninth century, the poet-prophet Ali bin Muhammad led imported East African slaves in Iraq during the Zanj Rebellion against the Abbasid Caliphate; Nanny of the Maroons was an 18th-century leader who rebelled against the British in Jamaica; and the Quilombo dos Palmares of Brazil flourished under Ganazumba (Ganga Zumba). The 1811 German Coast Uprising in the Territory of Orleans was the largest rebellion in the continental United States; Denmark Vesey rebelled in South Carolina, and Madison Washington during the Creole case in the 19th century the United States.
The Mamluk Sultanate reigned for centuries out of a slave rebellion in Egypt. It gave birth to both the Bahri dynasty and Burji dynasty and their countless artistic and scientific achievements. Among many accomplishments, the Mamluks were responsible for turning back the Mongol conquest.
Ancient Sparta had a special type of serf called helots who were often treated harshly, leading them to rebel. According to Herodotus (IX, 28–29), helots were seven times as numerous as Spartans. Every autumn, according to Plutarch (Life of Lycurgus, 28, 3–7), the Spartan ephors would pro forma declare war on the helot population so that any Spartan citizen could kill a helot without fear of blood or guilt in order to keep them in line (crypteia).
In the Roman Empire, though the heterogeneous nature of the slave population worked against a strong sense of solidarity, slave revolts did occur and were severely punished. The most famous slave rebellion in Europe was led by Spartacus in Roman Italy, the Third Servile War. This war resulted in the 6000 surviving rebel slaves being crucified along the main roads leading into Rome. This was the third in a series of unrelated Servile Wars fought by slaves against the Romans.
The English peasants' revolt of 1381 led to calls for the reform of feudalism in England and an increase in rights for serfs. The Peasants' Revolt was one of several popular revolts in late medieval Europe. Richard II agreed to reforms including fair rents and the abolition of serfdom. Following the collapse of the revolt, the king's concessions were quickly revoked, but the rebellion is significant because it marked the beginning of the end of serfdom in medieval England.
In Russia, the slaves were usually classified as kholops. A kholop's master had unlimited power over his life. Slavery remained a major institution in Russia until 1723, when Peter the Great converted the household slaves into house serfs. Russian agricultural slaves were formally converted into serfs earlier in 1679. During the 16th and 17th centuries, runaway serfs and kholops known as Cossacks, ("outlaws") formed autonomous communities in the southern steppes. There were numerous rebellions against slavery and serfdom, most often in conjunction with Cossack uprisings, such as the uprisings of Ivan Bolotnikov (1606–1607), Stenka Razin (1667–1671), Kondraty Bulavin (1707–1709), and Yemelyan Pugachev (1773–1775), often involving hundreds of thousands and sometimes millions. Between the end of the Pugachev rebellion and the beginning of the 19th century, there were hundreds of outbreaks across Russia.
Numerous African slave rebellions and insurrections took place in North America during the 17th, 18th, and 19th centuries. There is documented evidence of more than 250 uprisings or attempted uprisings involving 10 or more slaves. Three of the best known in the United States during the 19th century are the revolts by Gabriel Prosser in the Richmond, Virginia area in 1800, Denmark Vesey in Charleston, South Carolina, in 1822, and Nat Turner in Southampton County, Virginia, in 1831. Slave resistance in the antebellum South did not gain the attention of academic historians until the 1940s, when historian Herbert Aptheker started publishing the first serious scholarly work on the subject. Aptheker stressed how rebellions were rooted in the exploitative conditions of the southern slave system. He traversed libraries and archives throughout the South, managing to uncover roughly 250 similar instances.
Middle East
The Zanj Rebellion was the culmination of a series of small revolts. It took place near the city of Basra, in southern Iraq over fifteen years (869−883 AD). It grew to involve over 500,000 slaves, who were imported from across the Muslim empire.
The Mamluk Sultanate reigned for centuries out of a slave rebellion in Egypt. It gave birth to both the Bahri dynasty and Burji dynasty and their countless artistic and scientific achievements. Among many accomplishments, the Mamluks were responsible for turning back the Mongol conquest.
Europe and the Mediterranean
In the 3rd century BCE, Drimakos (or Drimachus) led a slave revolt on the slave entrepot of Chios, took to the hills and directed a band of runaways in operations against their ex-masters.
The Servile Wars (135 to 71 BCE) were a series of slave revolts within the Roman Republic.
The First and Second Servile War occurred in Sicily.
The Third Servile War (73 to 71 BCE) occurred in mainland Italy. Spartacus, an escaped gladiator supposedly from Thrace, became the most prominent of the rebel leaders; Marcus Licinius Crassus suppressed the insurgents. Many modern rebels (such as the Spartacus League have since regarded Spartacus as a heroic figure.
Other slave revolts occurred elsewhere.
Eumenes III, king of Pergamon from 133 to 129 BCE, promised freedom to slaves to draw support against the Roman Republic.
A number of slave revolts occurred in the Mediterranean area during the early modern period:
1748: Hungarian, Georgian and Maltese slaves on board the Ottoman ship Lupa revolted and sailed the ship to Malta.
1749: Conspiracy of the Slaves – Muslim slaves in Malta planned to rebel and take over the island, but plans leaked out beforehand and the would-be rebels were arrested and many were executed.
1760: Christian slaves on board the Ottoman ship Corona Ottomana revolted and sailed the ship to Malta.
São Tomé and Príncipe
On 9 July 1595, Rei Amador, and his people, the Angolars, allied with other enslaved Africans of its plantations, marched into the interior woods and battled against the Portuguese. It is said that day, Rei Amador and his followers raised a flag in front of the settlers and proclaimed Rei Amador as king of São Tomé and Príncipe, making himself as "Rei Amador, liberator of all the black people".
Between 1595 and 1596, part of the island of São Tomé was ruled by the Angolars, under the command of Rei Amador. On 4 January 1596, he was captured, sent to prison and was later executed by the Portuguese. Still today, they remember him fondly and consider him a national hero of the islands.
In the first decades of the 17th century, there were frequent slave revolts in the Portuguese colony of São Tomé and Príncipe, off the African shore, which damaged the sugar crop cultivation there.
South America and the Caribbean
December 25, 1522 rebellion in Diego Colón de Toledo's plantation in what is known today as Santo Domingo is the first known slave rebellion of the region.
Between 1538 and 1542, a Guaraní slave from present-day Paraguay named Juliana killed her Spanish master and urged other indigenous women to do the same, ending up executed by order of Álvar Núñez Cabeza de Vaca. Her rebellion is regarded as one of the earliest recorded indigenous uprisings against the Spanish colonization of the Americas.
Quilombo dos Palmares in Brazil, 1605 to 1694, led by Zumbi dos Palmarés.
San Basilio de Palenque in Colombia, 16th century to the present, led by Benkos Biohó.
St. John, 1733, in what was then the Danish West Indies. The St. John's Slave Rebellion is one of the earliest and longest lasting slave rebellions in the Americas. It ended with defeat, however, and many rebels, including one of the leaders Breffu, committed suicide rather than being recaptured.
The most successful slave uprising was the Haitian Revolution, which began in 1791 and was eventually led by Toussaint L'Ouverture, culminating in the independent black republic of Haiti.
Panama also has an extensive history of slave rebellions going back to the 16th century. Slaves were brought to the isthmus from many regions in Africa, including the modern day countries of the Congo, Senegal, Guinea, and Mozambique. Immediately before their arrival on shore, or very soon after, many enslaved Africans revolted against their captors or participated in mass maroonage or desertion. The freed Africans founded communities in the forests and mountains, organized guerrilla bands known as Cimarrones. They began a long guerrilla war against the Spanish Conquistadores, sometimes in conjunction with nearby indigenous communities like the Kuna and the Guaymí. Despite massacres by the Spanish, the rebels fought until the Spanish crown was forced to concede to treaties that granted the Africans a life without Spanish violence and incursions. The leaders of the guerrilla revolts included Felipillo, Bayano, Juan de Dioso, Domingo Congo, Antón Mandinga, and Luis de Mozambique.
In the 1730s, the militias of the Colony of Jamaica fought the Jamaican Maroons for a decade, before agreeing to sign peace treaties in 1739 and 1740, which recognised their freedom in five separate Maroon Towns.
Tacky's War (1760) was a slave uprising in Jamaica, which ran from May to July before it was put down by the British colonial government.
The Suriname slave rebellion was marked by constant guerrilla warfare by Maroons and in 1765-1793 by the Aluku. This rebellion was led by Boni.
The Berbice slave revolt in Guyana in 1763 was led by Cuffy.
Cuba had slave revolts in 1795, 1798, 1802, 1805, 1812 (the Aponte revolt), 1825, 1827, 1829, 1833, 1834, 1835, 1838, 1839–43 and 1844 (the La Escalera conspiracy and revolt).
Revolts on the Caribbean Islands
Vincent Brown, a professor of History and of African and African-American Studies at Harvard, has made a study of the Transatlantic Slave Trade. In 2013, Brown teamed up with Axis Maps to create an interactive map of Jamaican slave uprisings in the 18th century called, “Slave Revolt in Jamaica, 1760-1761, A Cartographic Narrative.” Brown's efforts have shown that the slave insurrection in Jamaica in 1760-61 was a carefully planned affair and not a spontaneous, chaotic eruption, as was often argued (due in large part to the lack of written records produced by the insurgents). Tacky's War was a widespread slave uprising across Jamaica in the 1760s.
Later, in 1795, several slave rebellions broke out across the Caribbean, influenced by the Haitian Revolution:
In Jamaica, the descendants of Africans who fought and escaped from slavery and established free communities in the mountainous interior of Jamaica (Maroons), fought to preserve their freedom from British colonialists, in what came to be known as the Second Maroon War. However, this featured just one of the five Maroon towns in Jamaica.
In Dominica there was the Colihault Uprising.
In Saint Lucia there was the Bush War.
In the Saint Vincent islands the Second Carib War broke out.
In Grenada there was the Fedon Rebellion.
Curaçao had a slave revolt in 1795, led by Tula.
In Venezuela, the insurrection led by José Leonardo Chirino occurred in 1795.
In Barbados, a slave revolt occurred in 1816, led by Bussa.
In Guyana there was the Demerara Rebellion of 1795.
In the British Virgin Islands, minor slave revolts occurred in 1790, 1823 and 1830.
In Cuba, there were several revolts starting in 1825 with an uprising in Guamacaro and ending with the revolts of 1843 in Matanzas. These revolts have been widely studied by scholars such as Robert L. Paquette, Gloria García, Manuel Barcia, Aisha K. Finch and Michele Reid-Vazquez.
In the Danish West Indies an 1848 slave revolt led to emancipation of all slaves in the Danish West Indies.
In Puerto Rico in 1821, Marcos Xiorro planned and conspired to lead a slave revolt against the sugar plantation owners and the Spanish Colonial government. Even though the conspiracy was unsuccessful, Xiorro achieved legendary status among the slaves and is part of Puerto Rico's folklore.
Brazil
Many slave rebellions occurred in Brazil, most famously the Bahia Rebellion of 1822-1830 and the Malê Revolt of 1835 by the predominantly Muslim West African slaves at the time. The term malê was commonly used to refer to Muslims at the time from the Yoruba word imale.
North America
Numerous slave rebellions and insurrections took place in North America during the 17th, 18th, and 19th centuries. There is documentary evidence of more than 250 uprisings or attempted uprisings involving ten or more slaves. One of the first was at San Miguel de Gualdape, the first European settlement in what would become the United States. Three of the best known in the United States during the 19th century are the revolts by Gabriel Prosser in Virginia in 1800, Denmark Vesey in Charleston, South Carolina in 1822, and Nat Turner's Slave Rebellion in Southampton County, Virginia, in 1831.
Drapetomania was a supposed mental illness invented by American physician Samuel A. Cartwright in 1851 that allegedly caused black slaves to run away. Today, drapetomania is considered an example of pseudoscience, and part of the edifice of scientific racism.
Slave resistance in the antebellum South did not gain the attention of academic historians until the 1940s, when historian Herbert Aptheker started publishing the first serious scholarly work on the subject. Aptheker stressed how rebellions were rooted in the exploitative conditions of the Southern slave system. He traversed libraries and archives throughout the South, managing to uncover roughly 250 similar instances.
The 1811 German Coast Uprising, which took place in rural southeast Louisiana, at that time the Territory of Orleans, early in 1811, involved up to 500 insurgent slaves. It was suppressed by local militias and a detachment of the United States Army. In retaliation for the deaths of two white men and the destruction of property, the authorities killed at least 40 black men in a violent confrontation (the numbers cited are inconsistent); at least 29 more were executed (combined figures from two jurisdictions, St. Charles Parish and Orleans Parish). There was a third jurisdiction for a tribunal and what amounted to summary judgments against the accused, St. John the Baptist Parish. Fewer than 20 men are said to have escaped; some of those were later caught and killed, on their way to freedom.
Although only involving about seventy slaves and free blacks, Turner's 1831 rebellion is considered to be a significant event in American history. The rebellion caused the slave-holding South to go into a panic. Fifty-five men, women, and children were killed, and enslaved blacks were freed on multiple plantations in Southampton County, Virginia, as Turner and his fellow rebels attacked the white institution of plantation slavery. Turner and the other rebels were eventually stopped by state militias (Aptheker, 1993). The rebellion resulted in the hanging of about 56 slaves, including Nat Turner himself. Up to 200 other blacks were killed during the hysteria which followed, few of whom likely had anything to do with the uprising. Fears afterwards led to new legislation passed by Southern states prohibiting the movement, assembly, and education of slaves, and reducing the rights of free people of color. In addition, the Virginia legislature considered abolishing slavery to prevent further rebellions. In a close vote, however, the state decided to keep slaves.
The abolitionist John Brown had already fought against pro-slavery forces in Bleeding Kansas for several years when he decided to lead a raid on a Federal arsenal in Harpers Ferry, Virginia. This raid was a joint attack by freed blacks and white men who had corresponded with slaves on plantations in order to create a general uprising among slaves. Brown carried hundreds of copies of the constitution for a new republic of former slaves in the Appalachians. But they were never distributed, and the slave uprisings that were to have helped Brown did not happen. Some believe that he knew the raid was doomed but went ahead anyway, because of the support for abolition it would (and did) generate. The U.S. military, led by Lieutenant Colonel Robert E. Lee, easily overwhelmed Brown's forces. But directly following this, slave disobedience and the number of runaways increased markedly in Virginia.
The historian Steven Hahn proposes that the self-organized involvement of slaves in the Union Army during the American Civil War composed a slave rebellion that dwarfed all others. Similarly, tens of thousands of slaves joined British forces or escaped to British lines during the American Revolution, sometimes using the disruption of war to gain freedom. For instance, when the British evacuated from Charleston and Savannah, they took 10,000 slaves with them. They also evacuated slaves from New York, taking more than 3,000 for resettlement to Nova Scotia, where they were recorded as Black Loyalists and given land grants.
North American slave revolts
Santo Domingo Slave Revolt (1521)
San Miguel de Gualdape (1526)
Bayano Wars (1548)
Gaspar Yanga's Revolt (c. 1570) near the Mexican city of Veracruz; the group escaped to the highlands and built a free colony
Revolt in Gloucester County, Virginia (1663)
New York Slave Revolt of 1712
Samba Rebellion (1731)
Stono Rebellion (1739)
New York Conspiracy of 1741 (alleged)
During the American Revolutionary War, slaves reacted to Dunmore's Proclamation and the Philipsburg Proclamation, fleeing and sometimes taking up arms in the British military against their former masters (for example in the Ethiopian Regiment).
Pointe Coupée Slave Conspiracy of 1791
Pointe Coupée Slave Conspiracy of 1795
Gabriel's Rebellion (1800)
Rebellions in a dozen North Carolina counties (May and June, 1802)
Chatham Manor Rebellion (1805)
Slaves in three North Carolina counties conspire to poison their owners, in some cases successfully. (1805)
1811 German Coast Uprising
Aponte Conspiracy (1812)
George Boxley Rebellion (1815)
Denmark Vesey's conspiracy (1822)
Nat Turner's slave rebellion (1831)
Baptist War (1831)
Black Seminole Slave Rebellion (1835–1838)
Amistad seizure (1839)
Creole case (1841) (the most successful slave revolt in US history)
1842 Slave Revolt in the Cherokee Nation
John Brown's raid on Harpers Ferry (1859) (failed attempt to organize a slave rebellion)
Africa
In 1808 and 1825, there were slave rebellions in the Cape Colony, newly acquired by the British. Although the slave trade was officially abolished in the British Empire by the Slave Trade Act of 1807, and slavery itself a generation later with the Slavery Abolition Act 1833, it took until 1850 to be halted in the territories which were to become South Africa.
Slave ship revolts
There are 485 recorded instances of slaves revolting on board slave ships. A few of these ships endured more than one uprising during their career.
Most accounts of revolts aboard slave ships are given by Europeans. There are few examples of accounts by slaves themselves. William Snelgrave reported that the slaves who revolted on the British ship Henry in 1721 claimed that those who had captured them were "Rogues to buy them" and that they were bent on regaining their liberty. Another example that Richardson gives is that of James Towne who gives the account of slaves stating that Europeans did not have the right to enslave and take them away from their homeland and "wives and children".
Richardson compares several factors that contributed to slave revolts on board ships: conditions on the ships, geographical location, and proximity to the shore. He suggests that revolts were more likely to occur when a ship was still in sight of the shore. The threat of attack from the shore by other Africans was also a concern. If the ship was hit by disease and a large portion of the crew had been killed, the chances of insurrection were higher. Where the slaves were captured also had an effect on the number of insurrections. In many places, such as the Bight of Benin and the Bight of Biafra, the percentage of revolts and the percentage of the slave trade match up. Yet ships taking slaves from Senegambia experienced 22 percent of shipboard revolts while only contributing to four and a half percent of the slave trade. Slaves coming from West Central Africa accounted for 44 percent of the trade while only experiencing 11 percent of total revolts.
Lorenzo J. Greene gives many accounts of slave revolts on ships coming out of New England. These ships belonged to Puritans who controlled much of the slave trade in New England. Most revolts on board ships were unsuccessful. The crews of these ships, while outnumbered, were disciplined, well fed, and armed with muskets, swords, and sometimes cannons, and they were always on guard for resistance. The slaves on the other hand were the opposite, armed only with bits of wood and the chains that bound them.
However, some captives were able to take over the ships that were their prisons and regain their freedom. On October 5, 1764 the New Hampshire ship Adventure captained by John Millar was successfully taken by its cargo. The slaves on board revolted while the ship was anchored off the coast and all but two of the crew, including Captain Millar, had succumbed to disease. Another successful slave revolt occurred six days after the ship Little George had left the Guinea coast. The ship carried ninety-six slaves, thirty-five of which were male. The slaves attacked in the early hours of the morning, easily overpowering the two men on guard. The slaves were able to load one of the cannons on board and fire it at the crew. After taking control of the ship they sailed it up the Sierra Leone River and escaped. After having defended themselves with muskets for several days below decks the crew lowered a small boat into the river to escape. After nine days of living on raw rice they were rescued.
Mariana P. Candido notes that enslaved Africans worked on the ships that transported other Africans into slavery. These men, 230 in all, were used onboard slave ships for their ability to communicate with the slaves being brought on board and to translate between Captain and slaver. Enslaved sailors were able to alleviate some of the fears that newly boarded slaves had, such as fear of being eaten. This was a double-edged sword. The enslaved sailors sometimes joined other slaves in the revolts against the captain they served. In 1812 enslaved sailors joined a revolt on board the Portuguese ship Feliz Eugenia just off the coast of Benguela. The revolt took place below decks. The sailors, along with many of the children who were on board, were able to escape using small boats.
See also
History of slavery
Labour revolt
List of revolutions and rebellions
Slave ship
Bibliography
Herbert Aptheker, American Negro Slave Revolts, 6. ed., New York: International Publ., 1993 - classic
Matt D. Childs, The 1812 Aponte Rebellion in Cuba and the Struggle Against African Slavery, Chapel Hill: University of North Carolina Press, 2006
David P. Geggus, ed., The Impact of the Haitian Revolution in the Atlantic World, Columbia: University of South Carolina Press, 2001
Eugene D. Genovese, From Rebellion to Revolution: Afro-American Slave Revolts in the Making of the Modern World, Louisiana State University Press 1980
Joao Jose Reis, Slave Rebellion in Brazil: The Muslim Uprising of 1835 in Bahia (Johns Hopkins Studies in Atlantic History and Culture), Johns Hopkins Univ Press 1993
Rodriguez, Junius P., ed. Encyclopedia of Slave Resistance and Rebellion. Westport, CT: Greenwood, 2007.
Rodriguez, Junius P., ed. Slavery in the United States: A Social, Political, and Historical Encyclopedia. Santa Barbara, CA: ABC-CLIO, 2007.
References and notes
Further reading
External links
Audio on slave revolts in the Caribbean
These maroons affiliated with Seminole Indians in Florida led a slave rebellion that would be the largest in U.S. history.
African-American history |
null | null | Star Wars: Episode III – Revenge of the Sith | eng_Latn | Star Wars: Episode III – Revenge of the Sith is a 2005 American epic space opera film written and directed by George Lucas. It stars Ewan McGregor, Natalie Portman, Hayden Christensen, Ian McDiarmid, Samuel L. Jackson, Christopher Lee, Anthony Daniels, Kenny Baker and Frank Oz. The sequel to The Phantom Menace (1999) and Attack of the Clones (2002), It is the sixth film in the Star Wars film series, the final installment in the Star Wars prequel trilogy, and third chronological chapter of the "Skywalker Saga".
Revenge of the Sith is set three years after the onset of the Clone Wars, established in Attack of the Clones. The Jedi are spread across the galaxy, leading a large-scale war against the Separatists. Following the death of Separatist leader Count Dooku, the Jedi Council dispatches Obi-Wan Kenobi on a mission to eliminate General Grievous, the head of the Separatist army, to put an end to the war. Meanwhile, after having visions of his wife Padmé Amidala dying in childbirth, Anakin Skywalker is tasked by the Council to spy on Palpatine, the Supreme Chancellor of the Galactic Republic and, secretly, a Sith Lord. Palpatine manipulates Anakin into turning to the dark side of the Force and becoming his apprentice, Darth Vader, with wide-ranging consequences for the galaxy.
Lucas began writing the script before production of Attack of the Clones ended, citing that he wanted the end of the trilogy to have similar aspects to a romantic tragedy, thus building into Darth Vader's state at the beginning of the next film. Production of Revenge of the Sith started in September 2003, and filming took place in Australia with additional locations in Thailand, Switzerland, China, Italy and the United Kingdom.
Revenge of the Sith premiered on May 15, 2005, at the Cannes Film Festival, then released worldwide on May 19, 2005. The film received positive reviews and has been deemed an improvement over the first two prequel films by many, with praise for its action sequences, mature themes, score, visual effects, darker tone, and the performances of McGregor, McDiarmid, Oz, and Jimmy Smits; criticism largely focused on the film's dialogue and aspects of Christensen's performance. It broke several box office records during its opening week and went on to earn over $868 million worldwide, making it the second-highest-grossing film in the Star Wars franchise at the time. It was the highest-grossing film in the U.S. and the second-highest-grossing film worldwide in 2005. The film also holds the record for the highest opening-day gross on a Thursday, making $50 million.
A miniseries set ten years after the events of the film, titled Obi-Wan Kenobi, is set to premiere on May 25, 2022 on Disney+, with McGregor and Christensen reprising their roles from the film.
Plot
Above Coruscant, Obi-Wan Kenobi and Anakin Skywalker lead a mission to rescue the kidnapped Supreme Chancellor Palpatine from the cyborg Separatist commander General Grievous. After infiltrating Grievous' flagship, Obi-Wan and Anakin battle the Sith Lord Count Dooku, whom Anakin overpowers and decapitates at Palpatine's urging. Grievous escapes the battle-torn ship, which the Jedi crash-land on Coruscant. There, Anakin reunites with his wife, Padmé Amidala, who reveals that she is pregnant. While initially excited, Anakin soon begins to have visions of Padmé dying in childbirth.
Palpatine appoints Anakin to the Jedi Council as his personal representative. The Council, suspicious of Palpatine, approves the appointment but declines to grant Anakin the rank of Jedi Master and instead instructs him to spy on Palpatine, diminishing Anakin's faith in the Jedi. Meanwhile, on Utapau, Grievous relocates the Separatist leaders to the volcanic planet Mustafar. Obi-Wan travels to Utapau to confront Grievous and kills him, while Yoda travels to the Wookiee homeworld of Kashyyyk to defend it from a Separatist invasion.
Palpatine tempts Anakin with his knowledge of the Force, including the dark side, and offers to teach him the power to prevent death in order to save Padmé's life. Anakin deduces that Palpatine is the Sith Lord behind the war and reports his treachery to Mace Windu, who confronts and subdues him. Desperate to save Padmé's life, Anakin severs Windu's hand before he can kill Palpatine, who kills Windu with Force lightning. Anakin pledges himself to the Sith, and Palpatine dubs him Darth Vader. Palpatine issues Order 66, which commands the clone troopers to kill their commanding Jedi generals across the galaxy, while Vader and a battalion of clone troopers kill the remaining Jedi in the Jedi Temple. Vader then travels to Mustafar to assassinate the Separatist leaders, while Palpatine declares himself Emperor before the Galactic Senate, transforming the Republic into the Galactic Empire, and denounces the Jedi as traitors.
Having survived the chaos, Obi-Wan and Yoda return to Coruscant, where Obi-Wan learns of Anakin's turn to the dark side. Yoda instructs Obi-Wan to confront Vader while he faces Palpatine. Obi-Wan seeks out Padmé to find out Vader's whereabouts and reveals his treachery. Padmé then travels to Mustafar – with Obi-Wan stowing away on her ship – and pleads with Vader to abandon the dark side and leave with her, but he refuses. Seeing Obi-Wan aboard the ship, and thinking that they are conspiring to kill him, Vader uses the Force to choke Padmé unconscious in a blind rage. Obi-Wan then engages Vader in a lightsaber duel that ends with Obi-Wan severing Vader's legs and left arm, leaving him at the bank of a lava flow. Obi-Wan picks up Vader's lightsaber, watches in horror as flames consume the fallen Jedi, and leaves him for dead.
On Coruscant, Yoda battles Palpatine until their duel reaches a stalemate. Yoda then flees with Senator Bail Organa and regroups with Obi-Wan and Padmé on the planetoid Polis Massa. There, Padmé gives birth to twins, whom she names Luke and Leia, and dies soon after, having lost her will to live. Palpatine retrieves a barely-alive Vader and takes him to Coruscant, where his mutilated body is treated and covered in a black armored suit. When Vader asks for Padmé, Palpatine says that she died in the heat of Vader's anger, and Vader screams in anguish.
Obi-Wan and Yoda plan to conceal the twins from the Sith and go into exile until the time is right to challenge the Empire. As Padmé's funeral takes place on her homeworld of Naboo, Palpatine and Vader supervise the construction of the Death Star. Bail takes Leia to his homeworld of Alderaan, where he and his wife adopt her, while Obi-Wan delivers Luke to his step-uncle and -aunt, Owen and Beru Lars, on Tatooine before going into exile to watch over the boy.
Cast
Ewan McGregor as Obi-Wan Kenobi, a Jedi Master, general of the Galactic Republic and Anakin Skywalker's mentor.
Natalie Portman as Padmé Amidala, a senator of Naboo who is secretly Anakin's wife and pregnant with their children.
Hayden Christensen as Anakin Skywalker / Darth Vader, a Jedi Knight, hero of the Clone Wars and former Padawan of Obi-Wan, who turns to the dark side of the Force and becomes a Sith Lord. Christensen also plays Vader in his suit.
Ian McDiarmid as Palpatine / Darth Sidious, the Supreme Chancellor of the Galactic Republic who is secretly a Sith Lord, and later the leader of the Galactic Empire. He takes advantage of Anakin's distrust of the Jedi and fear of Padmé dying to turn him towards the dark side, becoming Vader's master.
Samuel L. Jackson as Mace Windu, a Jedi Master and senior member of the Jedi Council.
Jimmy Smits as Bail Organa, a senator of Alderaan.
Christopher Lee as Count Dooku / Darth Tyranus, Sidious' Sith apprentice and the leader of the Separatists.
Anthony Daniels as C-3PO, Anakin and Padmé's personal protocol droid that Anakin created as a child.
Kenny Baker as R2-D2, Anakin's astromech droid.
Frank Oz as the voice of Yoda, a Jedi Grandmaster and the leader of the Jedi Council.
Peter Mayhew, Oliver Ford Davies, Ahmed Best, and Silas Carson reprise their roles as Chewbacca, Sio Bibble, Jar Jar Binks, and Nute Gunray and Ki-Adi-Mundi, respectively, from the previous films. Joel Edgerton and Bonnie Piesse also make cameo appearances, reprising their roles as Owen and Beru Lars respectively from Attack of the Clones. Sound engineer Matthew Wood provides the voice of General Grievous, the fearsome cyborg commander of the Separatists' droid army. Temuera Morrison portrays Commander Cody and the rest of the clone troopers. Bruce Spence portrays Tion Medon, local administrator of Utapau. Jeremy Bulloch (who played Boba Fett in The Empire Strikes Back and Return of the Jedi) appears as Captain Colton, the pilot of the Rebel Blockade Runner Tantive IV. Genevieve O'Reilly portrays senator Mon Mothma, though her speaking scene was ultimately cut. Rohan Nichol portrays Captain Raymus Antilles.
Wayne Pygram appears as a young Wilhuff Tarkin, and stunt coordinator Nick Gillard appears as a Jedi named Cin Drallig (his name spelled backward, without the 'k'). Editor Roger Barton's son Aidan Barton portrays Luke Skywalker and Leia Organa as infants. James Earl Jones makes an uncredited cameo appearance, reprising his role from previous films as the voice of Darth Vader.
Director and Star Wars creator George Lucas has a cameo as Baron Papanoida, a blue-faced alien in attendance at the Coruscant opera house. Lucas' son Jett portrays Zett Jukassa, a young Jedi-in-training. One of Lucas' daughters, Amanda, appears as Terr Taneel, seen in a security hologram, while his other daughter Katie plays a blue-skinned Pantoran named Chi Eekway, visible when Palpatine arrives at the Senate after being saved by the Jedi and talking to Baron Papanoida at the opera house. Christian Simpson appeared as a stunt double for Hayden Christensen.
Production
Writing
Lucas said he conceived the Star Wars saga's story in the form of a plot outline in 1973. However, he later clarified that, at the time of the saga's conception, he had not fully realized the details—only major plot points. The film's climactic duel has its basis in the Return of the Jedi novelization, in which Obi-Wan recounts his battle with Anakin that ended with the latter falling "into a molten pit". Lucas began working on the screenplay for Episode III before the previous film, Attack of the Clones, was released, proposing to concept artists that the film would open with a montage of seven battles on seven planets. In The Secret History of Star Wars, Michael Kaminski surmises that Lucas found flaws with Anakin's fall to the dark side and radically reorganized the plot. For example, instead of opening the film with a montage of Clone War battles, Lucas decided to focus on Anakin, ending the first act with him killing Count Dooku, an action that signals his turn to the dark side.
A significant number of fans speculated online about the episode title for the film with rumored titles including Rise of the Empire, The Creeping Fear (which was also named as the film's title on the official website on April Fool's 2004), and Birth of the Empire. Eventually, Revenge of the Sith also became a title guessed by fans that George Lucas would indirectly confirm. The title is a reference to Revenge of the Jedi, the original title of Return of the Jedi; Lucas changed the title scant weeks before the premiere of Return of the Jedi, declaring that Jedi do not seek revenge.
Since Lucas refocused the film on Anakin, he had to sacrifice certain extraneous plot points relating to Attack of the Clones. Lucas had previously promised fans that he would explain the mystery behind the erasure of the planet Kamino from the Jedi Archives. However, Lucas abandoned this plot thread in order to devote more time to Anakin's story, leaving the matter unresolved on film.
Lucas had originally planned to include even more ties to the original trilogy, and wrote early drafts of the script in which a 10-year-old Han Solo appeared on Kashyyyk, but the role was not cast or shot. He also wrote a scene in which Palpatine reveals to Anakin that he created him from midichlorians, and is thus his "father", a clear parallel to Vader's revelation to Luke in The Empire Strikes Back, but Lucas ejected this scene as well. Another planned scene by Lucas that was written during the early development of the film was a conversation between Master Yoda and the ghostly Qui-Gon Jinn, with Liam Neeson reprising his role as Jinn (he also hinted his possible appearance in the film). However, the scene was never filmed and Neeson was never recorded, although the scene was present in the film's novelization. Neeson finally reprised the role in an episode of The Clone Wars in 2011.
After principal photography was complete in 2003, Lucas made even more changes in Anakin's character, rewriting the entire "turn" sequence. Lucas accomplished this "rewrite" through editing the principal footage and filming new scenes during pickups in London in 2004. In the previous versions, Anakin had myriad reasons for turning to the dark side, one of which was his sincere belief that the Jedi were plotting to take over the Republic. Although this is still intact in the finished film, by revising and refilming many scenes, Lucas emphasized Anakin's desire to save Padmé from death. Thus, in the version that made it to theaters, Anakin falls to the dark side primarily to save Padmé.
Art design
For the Kashyyyk environment, the art department turned to the Star Wars Holiday Special for inspiration. Over a period of months, Lucas would approve hundreds of designs that would eventually appear in the film. He would later rewrite entire scenes and action sequences to correspond to certain designs he had chosen. The designs were then shipped to the pre-visualization department to create moving CGI versions known as animatics. Ben Burtt would edit these scenes with Lucas in order to pre-visualize what the film would look like before the scenes were filmed. The pre-visualization footage featured a basic raw CGI environment with equally unprocessed character models performing a scene, typically for action sequences. Steven Spielberg was brought in as a "guest director" to make suggestions to the art designers for the Mustafar duel, and oversee the pre-visualization of an unused version of the Utapau chase scene. Later, the pre-visualization and art department designs were sent to the production department to begin "bringing the film out of the concept phase" by building sets, props and costumes.
Filming
Although the first scene filmed was the final scene to appear in the film (shot during the filming of Attack of the Clones in 2000), the first bulk of principal photography on the film occurred from June 30, 2003 to September 17, 2003, with additional photography occurring at Shepperton Studios in Surrey and Elstree Studios in Hertfordshire from August 2004 to January 31, 2005. The initial filming took place on sound stages at Fox Studios Australia in Sydney, although practical environments were shot as background footage later to be composited into the film. These included the limestone mountains depicting Kashyyyk, which were filmed in Phuket, Thailand. The production company was also fortunate enough to be shooting at the same time that Mount Etna erupted in Italy. Camera crews were sent to the location to shoot several angles of the volcano that were later spliced into the background of the animatics and the final film version of the planet Mustafar.
While shooting key dramatic scenes, Lucas would often use an "A camera" and "B camera", or the "V technique", a process that involves shooting with two or more cameras at the same time in order to gain several angles of the same performance. Using the HD technology developed for the film, the filmmakers were able to send footage to the editors the same day it was shot, a process that would require a full 24 hours had it been shot on film. Footage featuring the planet Mustafar was given to editor Roger Barton, who was on location in Sydney cutting the climactic duel.
Hayden Christensen says Lucas asked him "to bulk up and physically show the maturity that had taken place between the two films." The actor says he worked out with a trainer in Sydney for three months and ate "six meals a day and on every protein, weight gain supplement that man has created" to go from 160 lbs to 185 lbs.
Christensen and Ewan McGregor began rehearsing their climactic lightsaber duel long before Lucas would shoot it. They trained extensively with stunt coordinator Nick Gillard to memorize and perform their duel together. As in the previous prequel film, McGregor and Christensen performed their own lightsaber fighting scenes without the use of stunt doubles. The speed at which Vader and Obi-Wan engage in their duel is mostly the speed at which it was filmed, although there are instances where single frames were removed to increase the velocity of particular strikes. An example of this occurs as Obi-Wan strikes down on Vader after applying an armlock in the duel's first half.
Revenge of the Sith eventually became the first Star Wars film in which Anakin Skywalker and the suited Darth Vader were played by the same actor in the same film. As Christensen recounted, it was originally intended to simply have a "tall guy" in the Darth Vader costume, but after "begging and pleading" Christensen persuaded Lucas to have the Vader costume used in the film created specifically to fit him. The new costume featured shoe lifts and a muscle suit. It also required Christensen (who is tall) to look through the helmet's mouthpiece.
In 2004, Gary Oldman was originally approached to provide the voice of General Grievous; however, complications arose during contract negotiations after Oldman learned the film was to be made outside of the Screen Actors Guild, of which he is a member. He backed out of the role rather than violate the union's rules. Matthew Wood, who voiced Grievous, disputed this story at Celebration III, held in Indianapolis. According to him, Oldman is a friend of producer Rick McCallum, and thus recorded an audition as a favor to him, but was not chosen. Wood, who was also the supervising sound editor, was in charge of the auditions and submitted his audition anonymously in the midst of 30 others, under the initials "A.S." for Alan Smithee. Days later, he received a phone call asking for the full name to the initials "A.S."
Visual effects
The post-production department (handled by Industrial Light & Magic) began work during filming and continued until weeks before the film was released in 2005. Special effects were created using almost all formats, including model work, CGI and practical effects. The same department later composited all such work into the filmed scenes—both processes taking nearly two years to complete. Revenge of the Sith has 2,151 shots that use special effects, a world record.
The DVD featurette Within a Minute illustrated the film required 910 artists and 70,441 man-hours to create 49 seconds of footage for the Mustafar duel alone. Members of Hyperspace, the Official Star Wars Fan Club, received a special look into the production. Benefits included not only special articles, but they also received access to a webcam that transmitted a new image every 20 seconds during the time it was operating in Fox Studios Australia. Many times the stars, and Lucas himself, were spotted on the webcam.
Deleted scenes
Lucas excised all scenes of a group of Senators, including Padmé, Bail Organa, and Mon Mothma (Genevieve O'Reilly), organizing an alliance to prevent Palpatine from usurping any more emergency powers. Though this is essentially the birth of the Rebel Alliance, the scenes were discarded to achieve more focus on Anakin's story. The scene where Yoda arrives on Dagobah to begin his self-imposed exile was also removed, but is featured as an extended scene in the DVD release, although McCallum stated he hoped Lucas would have added it to the new cut as part of a six-episode DVD box set.
Bai Ling filmed minor scenes for the film playing a senator, but her role was cut during editing. She claimed this was because she appeared in a nude pictorial for the June 2005 issue of Playboy, whose appearance on newsstands coincided with the film's May release. Lucas denied this, stating that the cut had been made more than a year earlier, and that he had cut his own daughter's scenes as well. The bonus features show an additional removed scene in which Jedi Master Shaak Ti is killed by General Grievous in front of Obi-Wan and Anakin. The bonus features also show Obi-Wan and Anakin running through Grievous' ship, escaping droids through a fuel tunnel, and arguing over what R2-D2 is saying.
Music
The music was composed and conducted by John Williams (who has composed and conducted the score for every episode in the Star Wars saga), and performed by the London Symphony Orchestra and London Voices. The film's soundtrack was released by Sony Pictures Classical Records on May 3, 2005, more than two weeks before the film's release. A music video titled A Hero Falls was created for the film's theme, "Battle of the Heroes", featuring footage from the film and was also available on the DVD.
The soundtrack also came with a collectors' DVD, Star Wars: A Musical Journey, at no additional cost. The DVD, hosted by McDiarmid, features 16 music videos set to remastered selections of music from all six film scores, set chronologically through the saga.
Releases
The first trailer premiered in theaters on November 5, 2004 with the release of The Incredibles. Just four months later, another trailer was unveiled on March 10, 2005, debuting with The O.C.s "The Mallpisode" during the second season and in theaters with the release of Robots the next day on March 11. Star Wars: Episode III – Revenge of the Sith charity premieres took place in Seattle, Los Angeles, Chicago, Washington D.C., Boston, Denver, Atlanta, San Francisco, and Miami on Thursday, May 12, 2005; and on May 13, 2005, there were two additional charity premiere screenings in George Lucas's hometown of Modesto, California. The official premiere was at the 2005 Cannes Film Festival (out of competition) on May 16. Its theatrical release in most other countries took place on May 19 to coincide with the 1999 release of The Phantom Menace (the 1977 release of A New Hope and the 1983 release of Return of the Jedi were also released on the same day and month, six years apart). The global outplacement firm Challenger, Gray & Christmas claimed one week before the premiere that it may have cost the U.S. economy approximately US$627 million in lost productivity because of employees who took a day off or reported in sick. Grauman's Chinese Theatre, a traditional venue for the Star Wars films, did not show it. However, a line of people stood there for more than a month hoping to convince someone to change this. Most of them took advantage of an offer to see the film at a nearby cinema, ArcLight Cinemas (formerly the "Cinerama Dome"). On May 16, the Empire Cinema in London's Leicester Square hosted a day-long Star Wars marathon showing of all six films; an army of Imperial stormtroopers "guarded" the area, and the Royal Philharmonic Orchestra gave a free concert of Star Wars music.
Leaked workprint
A copy of the film leaked onto peer-to-peer file sharing networks just hours after opening in theaters. The film was a time-stamped workprint, suggesting it may have come from within the industry rather than from someone who videotaped an advance screening. Eight people were later charged with copyright infringement and distributing material illegally. Documents filed by the Los Angeles District Attorney allege that a copy of the film was taken from an unnamed Californian post-production office by an employee, who later pleaded guilty to his charges. The illegal copy was passed among seven people until reaching an eighth party, who also pleaded guilty to uploading to an unnamed P2P network.
Shortly after the above-mentioned print was leaked, it was released in Shanghai as a bootleg DVD with Chinese subtitles. The unknown producer of this DVD also chose to include English subtitles, which were in fact translated back into English from the Chinese translation, rather than using the original English script. This resulted in the subtitles being heavily mis-translated, often causing unintentional humor. For example, in the opening crawl the title is mis-translated as Star War: The Third Gathers – The Backstroke of the West. Darth Vader's cry of "Noooooooo!" () is rendered as "Do not want". This translation error would later be popularized as an Internet meme. In 2016, YouTube user GratefulDeadpool uploaded a fan edit of Revenge of the Sith in which the cast's voices were dubbed by other actors to match the bootleg subtitles.
Rating
Revenge of the Sith is the first Star Wars film to receive a PG-13 rating from the Motion Picture Association of America (MPAA), officially for "sci-fi violence and some intense images", namely for the scene in which Darth Vader is set aflame by lava. Lucas had stated months before the MPAA's decision that he felt the film should receive a PG-13 rating, because of Anakin's final moments and the film's content being the darkest and most intense of all six films. Roger Ebert and Richard Roeper later opined that children would be able to handle the film as long as they had parental guidance. All previously released films in the series were rated PG.
Home media
The film was released on DVD and VHS on October 31, 2005, in the UK and Ireland; on November 1, 2005, it was released in the United States and Canada on DVD; and on November 3, 2005, it was released in Australia. It was also released in most major territories on or near the same day. The DVD release consists of widescreen and pan and scan fullscreen versions and is THX certified. This two-disc set contains one disc with the film and the other one with bonus features. The first disc features three randomized selected menus, which are Coruscant, Utapau and Mustafar. There is an Easter egg in the options menu. When the THX Optimizer is highlighted, the viewer can press 1-1-3-8. By doing this, a hip hop music video with Yoda and some clone troopers will play.
The DVD includes a number of documentaries including a new full-length documentary as well as two featurettes, one which explores the prophecy of Anakin Skywalker as the Chosen One, the other looking at the film's stunts and a of web-documentaries from the official web site. Like the other DVD releases, included is an audio commentary track featuring Lucas, producer Rick McCallum, animation director Rob Coleman, and ILM visual effects supervisors John Knoll and Roger Guyett. were included with introductions from Lucas and McCallum.
This release is notable because, due to marketing issues, it was the first Star Wars film never to be released on VHS in the United States. However, the film was released on VHS in Australia, the U.K. and other countries.
The DVD was re-released in a prequel trilogy box set on November 4, 2008.
The six Star Wars films were released by 20th Century Fox Home Entertainment on Blu-ray on September 16, 2011, in three different editions.
On April 7, 2015, Walt Disney Studios, 20th Century Fox, and Lucasfilm jointly announced the digital releases of the six released Star Wars films. Revenge of the Sith was released through the iTunes Store, Amazon Video, Vudu, Google Play, and Disney Movies Anywhere on April 10, 2015.
Walt Disney Studios Home Entertainment reissued Revenge of the Sith on Blu-ray, DVD, and digital download on September 22, 2019. Additionally, all six films were available for 4K HDR and Dolby Atmos streaming on Disney+ upon the service's launch on November 12, 2019. This version of the film was released by Disney on 4K Ultra HD Blu-ray on March 31, 2020 whilst being re-released on Blu-ray and DVD.
3D re-release
On September 28, 2010, it was announced that all six films in the series were to be stereo-converted to 3D. The films would be re-released in chronological order beginning with The Phantom Menace on February 10, 2012. Revenge of the Sith was originally scheduled to be re-released in 3D on October 11, 2013. However, on January 28, 2013, Lucasfilm announced that it was postponing the 3D release of episodes II and III in order to "focus 100 percent of our efforts on Star Wars: The Force Awakens" and that further information about 3D release plans would be issued at a later date. The premiere of the 3D version was shown on April 17, 2015, at Star Wars Celebration Anaheim.
Reception
Critical response
On review aggregator Rotten Tomatoes, the film has an approval rating of 80% based on 302 reviews, with an average rating of 7.3/10. The site's critical consensus reads, "With Episode III: Revenge of the Sith, George Lucas brings his second Star Wars trilogy to a suitably thrilling and often poignant – if still a bit uneven – conclusion." On Metacritic, the film has a weighted average score of 68 out of 100, based on 40 critics, indicating "generally favorable reviews". Audiences polled by CinemaScore gave the film an average grade of "A–" on an A+ to F scale, the same score as the previous two films.
Most critics have considered the film to be the best of the prequel trilogy. A. O. Scott of The New York Times concluded that it was "the best of the four episodes Mr. Lucas has directed", and equal to The Empire Strikes Back as "the richest and most challenging movie in the cycle". In a 2007 summary of the on Rotten Tomatoes, Revenge of the Sith was placed 51 out of 100, making it the only Star Wars prequel film to earn a rank on that list. J.R. Jones, a Chicago Reader critic who disliked The Phantom Menace and Attack of the Clones, gave the film a positive review, saying that it had a "relatively thoughtful story". Roger Ebert of the Chicago Sun-Times gave the film three-and-a-half out of four stars, writing "If [Lucas] got bogged down in solemnity and theory in Episode II: Attack of the Clones, the Force is in a jollier mood this time, and Revenge of the Sith is a great entertainment", but he noted that "the dialogue throughout the movie is once again its weakest point".
Though many critics and fans viewed Revenge of the Sith as the strongest of the three prequels, some viewers thought it was more or less on par with the previous two episodes. Much of the criticism was directed towards the dialogue, particularly the film's romantic scenes; critics claimed this demonstrated Lucas's weakness as a writer of dialogue, a subject with which Lucas openly agreed when receiving the Lifetime Achievement Award from the American Film Institute. Some film critics and fans criticized Hayden Christensen's acting, calling it "wooden".
In 2012, Glittering Images art critic Camille Paglia praised the film, comparing some of its scenes to works by modern painters and calling it "the greatest work of art in recent memory". Paglia explained that the film's finale
During the late 2010s, the film amassed a cult following on social media among young fans who were children when the film was released, using the film's dialogue to create Internet memes out of nostalgia.
Some American conservatives criticized the film, claiming it had a liberal bias and was a commentary on the George W. Bush administration and Iraq War. Some websites went so far as to propose a boycott of the film. Lucas defended the film, stating that the film's storyline was written during the Vietnam War and was influenced by that conflict rather than the war in Iraq. Lucas also said "The parallels between Vietnam and what we're doing in Iraq now are unbelievable".
Box office
The film was released in 115 countries. Its worldwide gross eventually reached $849 million—making it the second most financially successful film of 2005, behind Harry Potter and the Goblet of Fire. The film earned an estimated $16.91 million from 2,900 midnight screenings in North America upon its release. In total, it earned a record $50 million on its opening day. It was surpassed the following year by Pirates of the Caribbean: Dead Man's Chest, which earned $55.5 million on its opening day.
With only the May 19 earnings, the film broke four box office records: midnight screenings gross (previously held by The Lord of the Rings: The Return of the King, $8 million), opening day gross (Spider-Man 2, with $40.4 million), single day gross (Shrek 2 with $44.8 million) and Thursday gross (The Matrix Reloaded with $37.5 million). Its single day and opening day gross records were later surpassed by Pirates of the Caribbean: Dead Man's Chest on July 7, 2006, when that movie grossed $55.5 million on its opening day, and its midnight screening gross was broken by The Dark Knight on July 18, 2008 with $18.5 million.
According to box office analysis sites, the film set American records for highest gross in a given number of days for each of its first 12 days of release except for the seventh and eighth, where the record is narrowly held by Spider-Man 2. On its fifth day, it became the highest-grossing film of 2005, surpassing Hitch ($177.6 million). The film earned $158.5 million in its first four-day period, surpassing the previous four-day record held by The Matrix Reloaded ($134.3 million), and joining Spider-Man, The Matrix Reloaded and Harry Potter and the Goblet of Fire as one of the only films to make $100 million in three days. In eight days, it reached the $200 million mark (a record tied with Spider-Man 2) and by its 17th day, the film had passed $300 million (surpassing the record of 18 days of Shrek 2). It was eventually the third-fastest film (after Shrek 2 and Spider-Man) to reach $350 million.
The film ended its run in American theaters on October 20, 2005, finishing with a total gross of $380,270,577. It ranks 29th in all-time domestic grosses and is the highest-grossing U.S. of 2005, out-grossing second-place The Chronicles of Narnia: The Lion, the Witch and the Wardrobe by nearly $90 million. The film sold an estimated 59,324,600 tickets in the US. It topped the domestic box office for two consecutive weekends before being overtaken by Madagascar and The Longest Yard (which were in their second weekends).
International grosses that exceeded $460 million include those Australia ($27.2 million), France and Algeria ($56.9 million), Germany ($47.3 million), Italy ($11.3 million), Japan ($82.7 million), Mexico ($15.3 million), South Korea ($10.3 million), Spain ($23.8 million), and the United Kingdom and Ireland ($72.8 million).
Accolades
Following the release of Revenge of the Sith—the completion of the original and prequel Star Wars series—on June 9, 2005, George Lucas was presented with the 33rd American Film Institute Lifetime Achievement Award. The institute honored his "astonishing contributions to the art and technology of filmmaking, as well as the impact of the epic Star Wars series".
Despite being the prequel trilogy's best reviewed and received film, it received fewer award nominations than the previous films. It became the only Star Wars film not to be nominated for an Academy Award for Best Visual Effects; it was nominated for Best Makeup (Dave Elsey and Nikki Gooley), losing to The Chronicles of Narnia: The Lion, the Witch and the Wardrobe. It also won "Favorite Motion Picture" and "Favorite Dramatic Motion Picture" awards at the People's Choice Awards, "Hollywood Movie of the Year" award at the Hollywood Film Festival, Empire Awards (Sci-Fi/Fantasy Film), and the Teen Choice Award for Choice Movie – Action.
As did every film of the original trilogy, the film won the Saturn Award for Best Science Fiction Film. Williams also won Best Music. The film was nominated for ten Saturn Awards overall, including Best Director and Best Writing for Lucas, Best Actor for Christensen, Best Actress for Natalie Portman and Best Supporting Actor for Ian McDiarmid.
Of the three Star Wars prequels, the film received the fewest Golden Raspberry Awards nominations: only one, for Christensen as Worst Supporting Actor, which he won. (The Phantom Menace and Attack of the Clones received seven nominations each, with one and two wins, respectively.) It is the only Star Wars prequel not to receive a Razzie nomination for "Worst Picture". Christensen further won the "Best Villain" award at the MTV Movie Awards. The film also received the fewest nominations (and no wins) at the 2005 Stinkers Bad Movie Awards: Worst Screenplay for a Film Grossing More Than $100M and Worst On-Screen Couple (Christensen and Portman).
Themes
Throughout Revenge of the Sith, Lucas refers to a wide range of films and other sources, drawing on political, military, and mythological motifs to enhance his story's impact. The most media coverage was likely given to an exchange between Anakin and Obi-Wan, leading to the aforementioned conflict: "If you're not with me, then you're my enemy", Anakin declares. Despite Lucas' insistence to the contrary, The Seattle Times concluded, "Without naming Bush or the Patriot Act, it's all unmistakable no matter what your own politics may be."
McDiarmid, Lucas, and others have also called Anakin's journey to the dark side Faustian in the sense of making a "pact with the devil" for short-term gain, with the fiery volcano planet Mustafar representing hell. Midway through the film, Lucas intercuts between Anakin and Padmé by themselves, thinking about one another in the Jedi Temple and their apartment, respectively, during sunset. The sequence is without dialogue and complemented by a moody, synthesized soundtrack. Lucas' coverage of the exterior cityscapes, skylines and interior isolation in the so-called "Ruminations" sequence is similar to the cinematography and mise-en-scène of Rosemary's Baby, a film in which a husband makes a literal pact with the devil.
Other media
Novelization
The film's novelization was written by Matthew Stover. It has more dialogue than the film, and certain story elements were expanded upon in the novelization including Anakin and Palpatine's relationship and Palpatine's apprenticeship to Darth Plagueis.
Video game
A video game based on the film was released on May 5, 2005, two weeks before the film. The game generally followed the film's storyline, integrating scenes from the film. However, many sections of the game featured scenes cut from the film, or entirely new scenes for the game. The style of the game was mostly lightsaber combat and fighting as Obi-Wan or Anakin. It also has a form of multiplayer mode, which includes both "VS" and "Cooperative" mode. In the first mode, two players fight with characters of their choice against each other in a lightsaber duel to the death. In the latter mode, two players team up to combat increasingly difficult waves of enemies.
The Clone Wars
The 2008 animated film and subsequent television series fill the three-year gap between the events of Attack of the Clones and Revenge of the Sith. A number of plot threads initially developed for inclusion in Revenge of the Sith were instead incorporated into The Clone Wars. These include Boba Fett's revenge plot against Mace Windu for his father Jango's death, and the solving of the mystery behind deceased Jedi Master Sifo-Dyas which was introduced in Attack of the Clones. The final four episodes of the series take place concurrently with Revenge of the Sith. Several scenes from the film were recreated and expanded for these episodes in order to showcase the whereabouts of Anakin Skywalker's former Padawan Ahsoka Tano during the events of the film. While Ahsoka was a major character in The Clone Wars, she is not referenced in Revenge of the Sith; the character had not yet been created at the time.
References
Footnotes
Citations
Sources
External links
at
at
2000s adventure films
2000s English-language films
2000s pregnancy films
2000s science fiction adventure films
2005 films
2005 science fiction action films
2005 science fiction films
20th Century Fox films
American films about revenge
American films
American pregnancy films
American science fiction war films
American sequel films
Film and television memes
Films directed by George Lucas
Films produced by Rick McCallum
Films scored by John Williams
Films shot at Elstree Film Studios
Films shot at Shepperton Studios
Films shot in China
Films shot in Italy
Films shot in Sydney
Films with screenplays by George Lucas
Films shot in Switzerland
Films shot in Thailand
Interquel films
Lucasfilm films
Prequel films
Revenge of the Sith
Films about coups d'état |
null | null | Liturgical year | eng_Latn | The liturgical year, also known as the church year or Christian year, as well as the kalendar, consists of the cycle of liturgical seasons in Christian churches that determines when feast days, including celebrations of saints, are to be observed, and which portions of Scripture are to be read either in an annual cycle or in a cycle of several years.
Distinct liturgical colours may be used in connection with different seasons of the liturgical year. The dates of the festivals vary somewhat among the different churches, although the sequence and logic is largely the same.
Liturgical cycle
The liturgical cycle divides the year into a series of seasons, each with their own mood, theological emphases, and modes of prayer, which can be signified by different ways of decorating churches, colours of paraments and vestments for clergy, scriptural readings, themes for preaching and even different traditions and practices often observed personally or in the home. In churches that follow the liturgical year, the scripture passages for each Sunday (and even each day of the year in some traditions) are specified in a lectionary. After the Protestant Reformation, Anglicans and Lutherans continued to follow the lectionary of the Roman Rite. Following a decision of the Second Vatican Council, the Catholic Church revised that lectionary in 1969, adopting a three-year cycle of readings for Sundays and a two-year cycle for weekdays.
Adaptations of the revised Roman Rite lectionary were adopted by Protestants, leading to the publication in 1994 of the Revised Common Lectionary for Sundays and major feasts, which is now used by many Protestant denominations, including also Methodists, Reformed, United, etc. This has led to a greater awareness of the traditional Christian year among Protestants, especially among mainline denominations.
Biblical calendars
Scholars are not in agreement about whether the calendars used by the Jews before the Babylonian exile were solar (based on the return of the same relative position between the sun and the earth), lunisolar (based on months that corresponded to the cycle of the moon, with periodic additional months to bring the calendar back into agreement with the solar cycle) like the present-day Jewish calendar of Hillel II, or lunar, such as the Hijri calendar.
The first month of the Hebrew year was called אביב (Aviv), evidently adopted by Moses from Ipip as the eleventh month of the non-lunar Egyptian calendar (that is also the origin of Abib as the tenth month of the non-lunar Ethiopian calendar), meaning the month of green ears of grain. Having to occur at the appropriate time in the spring, it thus was originally part of a tropical calendar. At about the time of the Babylonian exile, when using the Babylonian civil calendar, the Jews adopted the term ניסן (Nisan) as the name for the month, based on the Babylonian name Nisanu. Thomas J Talley says that the adoption of the Babylonian term occurred even before the exile.
In the earlier calendar, most of the months were simply called by a number (such as "the fifth month"). The Babylonian-derived names of the month that are used by Jews are:
Nisan (March–April)
Iyar (April–May)
Sivan (May–June)
Tammuz (June–July)
Av (July–August)
Elul (August–September)
Tishrei (September–October)
Marcheshvan (October–November)
Kislev (November–December)
Tevet (December–January)
Shevat (January–February)
Adar 1 (February; only during leap years)
Adar (February–March)
In Biblical times, the following Jewish religious feasts were celebrated:
Pesach (Passover) – 14 Nisan (sacrifice of a lamb), 15 Nisan (Passover seder)
Chag HaMatzot (Unleavened Bread) – 15–21 Nisan
Reishit Katzir (Firstfruits) – 16 Nisan
Shavuot (Weeks) – Fiftieth day counted from Passover, normally 6–7 Sivan
Rosh Hashanah (Trumpets) – 1–2 Tishrei
Yom Kippur (Atonement) – 10 Tishrei
Sukkot (Tabernacles) – 15–21 Tishrei
Chanukah (Dedication) – 25 Kislev–2/3 Tevet (instituted in 164 BC)
Purim (Lots) – 14–15 Adar (instituted in c. 400 BC)
Eastern Christianity
East Syriac Rite
The Liturgical Calendar of the East Syriac Rite is fixed according to the flow of salvation history. With a focus upon the historical life of Jesus Christ, believers are led to the eschatological fulfillment (i.e. the heavenly bliss) through this special arrangement of liturgical seasons. The liturgical year is divided into 8 seasons of approximately 7 weeks each but adjusted to fit the solar calendar. The arrangement of the Seasons in the Liturgical Year is based on seven central events on celebrations of the Salvation History. They are:
Nativity of Christ
Epiphany of Christ
Resurrection of Christ
Pentecost
Transfiguration
Glorious Cross
Parousia (the Dedication of Church after Christ's second coming)
One of the oldest available records mentioning the liturgical cycle of east-syriac rite is handwritten manuscript named 'Preface to Hudra' written by Rabban Brick-Iso in 14th century. The manuscript mentions that the liturgical year is divided into nine seasons starting from Subara and ends with Qudas Edta. Catholic churches of east-syriac rite maintains the same liturgical calender till date except that many consider 7th and 8th seasons as a single one. The biblical reading and prayers during Mass and Liturgy of the Hours vary according to different seasons in the liturgical calendar.
Liturgical Calendar
The various seasons of the liturgical calendar of Syro-Malabar Church and Chaldean Catholic Church are given below.
Annunciation (Subara)
Weeks of Annunciation (Subara) is the first season of the liturgical year. The liturgical year begins with the commemoration of biblical events leading to the annunciation and birth of Jesus as expected savior in the old testament. The season begins on the Sunday just before the first of December and ends with the feast of Epiphany that is the Feast of the Baptism of Jesus. The faithful practice abstinence during December 1–25 in preparation for Christmas; this period is called "25 days Lent".
Feasts celebrated during this season
Feast of the Immaculate Conception of Mary, mother of Jesus (December 8)
Feast of Miraculous Cross of Mylapore (Saint Thomas Christian cross) (December 18) in Syro Malabar Church
Nativity of Our Lord and Saviour Jesus Christ or Christmas (December 25)
Feast of Holy Infants (December 28)
Feast of Name Iso (January 1)
Feast of Mary, mother of Jesus (last Friday of Season)
Epiphany (Denha)
Weeks of Epiphany begins on the Sunday closest to the feast of Epiphany and runs to the beginning of Great Fast. The word denha in Syriac means sunrise. Church considers the baptism of Jesus in Jordan as the first historical event when the Holy Trinity is revealed to humans in the humanity of Jesus Christ. Thus the season commemorates the manifestation or revelation of Jesus and Trinity to the world. During the season the church celebrates the feasts of Saints in connection with the manifestation of the Lord.
Feasts celebrated during the period
Feast of Epiphany or Feast of Baptism of the Lord (January 6)
Feast of Saint John the Baptist on first Friday of Epiphany
Feast of Apostles Peter (Kepha) and Paul on second Friday of Epiphany
Feast of Evangelists on third Friday of Epiphany
Feast of Saint Stephan on fourth Friday of Epiphany
Feast of Fathers of Church on fifth Friday of Epiphany
Feast of Patron Saint of Church on sixth Friday of Epiphany
Feast of all departed faithful on last Friday of Epiphany
Great Fast (Sawma Rabba)
During these weeks the faithful meditate over the 40 days fast of Jesus and the culmination of his public life in passion, death and burial. The season begins 50 days before Easter on Peturta Sunday and comprises the whole period of Great Lent and culminates on Resurrection Sunday. Word Peturta in Syriac means "looking back" or "reconciliation".
Faithful enter the weeks of Great Fast, celebrating the memory of all the Faithful Departed on the last Friday of Denha. According to the ecclesial tradition, the weeks of Great Fast is also an occasion to keep up the memory of the beloved Departed through special prayers, renunciation, almsgiving, and so on and thus prepare oneself for a good death and resurrection in Jesus Christ. During the fast faithful of Syro Malabar Church do not use meat, fish, egg, many dairy products, and most favorite food items, and also avoid sexual contacts on all days including Sundays and Feast days. Before European colonization, Indian Nasranis used to have food only once a day (after 3:00 pm) on all days during Great Fast.
-
Feasts in the Lenten Season
Peturta Sunday on First Sunday of Great Fast
Ash Monday or Clean Monday on the first day (Monday) of Great Fast
Lazarus Friday on the sixth Friday of Great Fast
Oshana Sunday on the seventh Sunday of Great Fast
Thursday of Pesha
Friday of Passion or Good Friday
Great Saturday or Saturday of Light
The following feasts are always in the Lenten Season:
Feast of Mar Cyril of Jerusalem (March 18)
Feast of Saint Joseph (March 19)
Feast of the Annunciation (March 25)
Resurrection (Qyamta)
The weeks of Great Resurrection begin on the Resurrection Sunday and run to the feast of Pentecost. The Church celebrates the Resurrection of our Lord during these seven weeks: Jesus’ victory over death, sin, suffering and Satan. The church also commemorates various events that occurred after the resurrection of Christ, such as the visits of Jesus to the Apostles and the ascension of Jesus. According to eastern Christianity, the Feast of Resurrection is the most important and the greatest feast in a liturgical year. Therefore, the season commemorating the resurrection of Christ is also of prime importance in the church liturgy. The first week of the season is celebrated as the 'Week of weeks' as it is the week of the resurrection of Christ.
Feasts celebrated during the period:
Feast of Resurrection of Christ
Feast of All Confessors (Saints) on the first Friday of Qyamta
New Sunday or St. Thomas Sunday on the second Sunday of Qyamta
Feast of Ascension of Jesus on the sixth Friday of Qyamta
The following feasts are always in the season of resurrection:
Feast of Saint George (April 24)
Feast of Mark the Evangelist (April 25)
Feast of Saint Joseph, the worker (May 1)
Feasts of Saint Philip and Saint James the apostles (May 3)
Apostles (Slihe)
Weeks of apostles (Slihe) starts on the feast of Pentecost, fiftieth day of the Resurrection Sunday. During these days the church commemorates the inauguration of church and the acts of apostles and church fathers through which the foundation of the church was laid. Church meditates on the virtues of the early church: fellowship, breaking of bread and sharing of wealth, and the fruits and gifts of Holy Spirit. The spread of the church all over the world as well as her growth is also remembered during this season.
Feast celebrated during the season:
Feast of Pentecost on first Sunday of Slihe
Feast of Friday of Gold: The first commemoration of the first miracle of apostles done by Saint Peter.
The following feasts are commemorated in the season of Slihe
Feast of Mar Aphrem (June 10)
Feast of the Apostles Peter and Paul (June 29)
Feast of Mar Thoma, founding father of east Syriac churches (July 3)
Feast of Mar Quriaqos and Yolitha (July 15)
Qaita (Summer)
Along the weeks of Qaita maturity and fruitfulness of church are commemorated. The Syriac word Qaita means “summer” and it is a time of harvest for the Church. The fruits of the Church are those of holiness and martyrdom. While the sprouting and infancy of the Church were celebrated in 'the Weeks of the Apostles,' her development in different parts of the world by reflecting the image of the heavenly Kingdom and giving birth to many saints and martyrs are proclaimed during this season. Fridays of this Season are set apart for honoring Saints and Martyrs.
Feast celebrated during the season:
Feast of the twelve apostles and Nusardeil on the first Sunday of Qaita (Nusardeil is a Persian word which means "God-given New Year Day").
Feast of Mar Jacob of Nisibis on the first Friday of Qaita.
Feast of Mar Mari on the second Friday of Qaita.
Feast of Marta Simoni and her Seven Children on the fifth Friday of Qaita.
Feast of Mar Shimun Bar Sabbai and Companions on the sixth Friday of Qaita.
Feast of martyr Mar Quardag on the seventh Friday of Qaita.
The following feasts are commemorated in the season of Qaita
Feast of seventy disciples of Jesus (July 27)
Feast of Saint Alphonsa in Syro Malabar Catholic Church (July 28)
Feast of Transfiguration of Jesus (August 6)
Feast of Assumption of Mary (August 15)
Eliyah-Sliba-Moses
The name of the seasons of Eliyah-Sliba-Moses takes their origin from the feast of the transfiguration of Jesus. And the seasons revolve around the exaltation of the cross on the feast of the glorious cross on September 14. During the seasons of Eliyah and Sliba church reminds the faithful of the heavenly bliss which is promised to be inherited at the end of earthly life and the church commemorates the exaltic experience of the bliss through various sacraments. While during the season of Moses church meditates upon the end of time and the last judgment. Many at times the season of Moses is regarded as a distinct and separate season from the other two since it has a distinct theme.
The season of Eliyah has a length of one to three Sundays. Season of Sliba starts on Sunday on or after the feast of the glorious cross and has a length of three to four weeks. The first Sunday of Sliba is always considered as the fourth Sunday of the combined season. The season of Moses always has four weeks.
Feast celebrated during the seasons:
Feast of the glorious Cross
The following feasts are commemorated in the seasons of Eliyah-Sliba-Moses
Feast of Nativity of Mary on September 8 and the eight-day fast in preparation for the feast
Dedication of the church (Qudas Edta)
The weeks of the dedication of the church is the last liturgical season in the East Syriac rite. It consists of four weeks and ends on the Saturday before Sunday between November 27 and December 3. The theme of the season is that the church is presented by Christ as his eternal bride before his father at the heavenly bride chamber. The period has its origin in the feast of the dedication of the church of Sephelcure or the Jewish feast of Hanukkah. However, the season was officially instituted by Patriarch Isho-Yahb III of Seleucia-Ctesiphon (647–657) by separating it from the season of Moses.
Feasts celebrated during the season:
Feast of dedication of the church on 1st Sunday of Qudas Edta
Feast of Christ the King on last Sunday of Qudas Edta (Celebrated only in eastern catholic churches of the rite since pope Pius XI instituted it in Roman-rite).
Eastern Orthodox Church
The liturgical year in the Eastern Orthodox Church is characterized by alternating fasts and feasts, and is in many ways similar to the Catholic year. However, Church New Year (Indiction) traditionally begins on September 1 (Old Style or New Style), rather than the first Sunday of Advent. It includes both feasts on the Fixed Cycle and the Paschal Cycle (or Moveable Cycle). The most important feast day by far is the Feast of Pascha (Easter) – the Feast of Feasts. Then the Twelve Great Feasts, which commemorate various significant events in the lives of Jesus Christ and of the Theotokos (Virgin Mary).
The majority of Orthodox Christians (Russians, in particular) follow the Julian Calendar in calculating their ecclesiastical feasts, but many (including the Ecumenical Patriarchate and the Church of Greece), while preserving the Julian calculation for feasts on the Paschal Cycle, have adopted the Revised Julian Calendar (at present coinciding with the Gregorian Calendar) to calculate those feasts which are fixed according to the calendar date.
Between 1900 and 2100, there is a thirteen-day difference between the dates of the Julian and the Revised Julian and Gregorian calendars. Thus, for example, where Christmas is celebrated on December 25 O.S. (Old Style), the celebration coincides with January 7 in the Revised Calendar. The computation of the day of Pascha (Easter) is, however, always computed according to a lunar calendar based on the Julian Calendar, even by those churches which observe the Revised Calendar.
There are four fasting seasons during the year: The most important fast is Great Lent which is an intense time of fasting, almsgiving and prayer, extending for forty days prior to Palm Sunday and Holy Week, as a preparation for Pascha. The Nativity Fast (Winter Lent) is a time of preparation for the Feast of the Nativity of Christ (Christmas), but whereas Advent in the West lasts only four weeks, Nativity Fast lasts a full forty days. The Apostles' Fast is variable in length, lasting anywhere from eight days to six weeks, in preparation for the Feast of Saints Peter and Paul (June 29). The Dormition Fast lasts for two weeks from August 1 to August 14 in preparation for the Feast of the Dormition of the Theotokos (August 15). The liturgical year is so constructed that during each of these fasting seasons, one of the Great Feasts occurs, so that fasting may be tempered with joy.
In addition to these fasting seasons, Orthodox Christians fast on Wednesdays and Fridays throughout the year (and some Orthodox monasteries also observe Monday as a fast day). Certain fixed days are always fast days, even if they fall on a Saturday or Sunday (in which case the fast is lessened somewhat, but not abrogated altogether); these are: The Decollation of St. John the Baptist, the Exaltation of the Cross and the day before the Epiphany (January 5). There are several fast-free periods, when it is forbidden to fast, even on Wednesday and Friday. These are: the week following Pascha, the week following Pentecost, the period from the Nativity of Christ until January the 5th and the first week of the Triodion (the week following the 17th Sunday before Pentecost).
Pascha
The greatest feast is Pascha. Easter for both East and West is calculated as the first Sunday after the full moon that falls on or after March 21 (nominally the day of the vernal equinox), but the Orthodox calculations are based on the Julian calendar, whose March 21 corresponds at present with April 3 of the Gregorian calendar, and on calculations of the date of full moon different from those used in the West (see computus for further details).
The date of Pascha is central to the entire ecclesiastical year, determining not only the date for the beginning of Great Lent and Pentecost, but affecting the cycle of moveable feasts, of scriptural readings and the Octoechos (texts chanted according to the eight ecclesiastical modes) throughout the year. There are also a number of lesser feasts throughout the year that are based upon the date of Pascha. The moveable cycle begins on the Zacchaeus Sunday (the first Sunday in preparation for Great Lent or the 33rd Sunday after Pentecost as it is known), though the cycle of the Octoechos continues until Palm Sunday.
The date of Pascha affects the following liturgical seasons:
The period of the Triodion (the Sundays before Great Lent, Cheesefare Week, Palm Sunday, and Holy Week)
The period of the Pentecostarion (Sunday of Pascha through the Sunday After Pentecost which is also called the Sunday of all saints)
The twelve Great Feasts
Some of these feasts follow the Fixed Cycle, and some follow the Moveable (Paschal) Cycle. Most of those on the Fixed Cycle have a period of preparation called a Forefeast, and a period of celebration afterward, similar to the Western Octave, called an Afterfeast. Great Feasts on the Paschal Cycle do not have Forefeasts. The lengths of Forefeasts and Afterfeasts vary, according to the feast.
Nativity of the Theotokos (September 8)
birth of the Theotokos to Joachim and Anna
Elevation of the Cross (September 14)
the rediscovery of the original Cross on which Christ was crucified
Entrance of the Theotokos into the Temple (November 21)
the entry of the Theotokos into the Temple around the age of 3
Nativity of Our Lord and Saviour Jesus Christ (December 25)
the birth of Jesus, or Christmas
Theophany (January 6)
the baptism of Jesus Christ, Christ's blessing of the water, and the revealing of Christ as God
Presentation of Our Lord in the Temple (February 2)
Christ's presentation as an infant in the Temple by the Theotokos and Joseph.
Annunciation of the Theotokos (March 25)
Gabriel's announcement to the Theotokos that she will conceive the Christ, and her wilful agreement thereto
Note: In Eastern practice, should this feast fall during Holy Week or on Pascha itself, the feast of the Annunciation is not transferred to another day. In fact, the conjunction of the feasts of the Annunciation and Pascha (dipli Paschalia, ) is considered an extremely festive event.
Entry into Jerusalem (Sunday before Pascha)
known in the West as Palm Sunday.
Ascension (40 days after Pascha)
Christ's ascension into Heaven following his resurrection.
Pentecost (50 days after Pascha)
The Holy Spirit comes and indwells the apostles and other Christian believers.
Transfiguration of Our Lord (August 6)
Christ's Transfiguration as witnessed by Peter, James and John.
Dormition of the Theotokos (August 15)
The falling asleep of the Theotokos (cf. the Assumption of Mary in Western Christianity)
Other feasts
Some additional feasts are observed with as though they were Great Feasts:
The Protection of the Mother of God (October 1), especially among the Russian Orthodox
The Feast of Saint James the Just (October 23)
The Feast of Saint Demetrius of Thessaloniki (October 26)
The Feast of the Holy Archangels Michael and Gabriel (November 8)
The Feast of Saint Nicholas, the Bishop of Myra in Lycia (December 6)
The Feast of the Conception of Mary by Saints Joachim and Anne (December 9)
The Feast of Saint Spiridon (December 12)
The Feast of Saint Stephen the Deacon (December 27)
The Feast of Saint Basil the Great and the Circumcision of Christ (January 1)
The Feast of the Three Holy Hierarchs: Basil the Great, Gregory the Theologian and John Chrysostom (January 30)
The Feast of the Forty Martyrs of Sebaste (March 9)
The Feast of Saint Patrick (March 17)
The Feast of Saint George (April 23)
The Feast of the Holy Emperors Constantine and Helen (May 21)
The Nativity of Saint John the Baptist (June 24)
The Feast of Saints Peter and Paul (June 29)
The Feast of Saint Elijah the Prophet (July 20)
The Feast of Saint Christina of Bolsena the Great Martyr (July 24)
The Beheading of St. John the Baptist (August 29)
Beginning of the Indiction-Ecclesiastical Year (September 1)
The Patronal Feast of a church or monastery
Every day throughout the year commemorates some saint or some event in the lives of Christ or the Theotokos. When a feast on the moveable cycle occurs, the feast on the fixed cycle that was set for that calendar day is transferred, with the propers of the feast often being chanted at Compline on the nearest convenient day.
Cycles
In addition to the Fixed and Moveable Cycles, there are a number of other liturgical cycles in the ecclesiastical year that affect the celebration of the divine services. These include, the Daily Cycle, the Weekly Cycle, the Cycle of Matins Gospels, and the Octoechos.
Oriental Orthodox and P'ent'ay Evangelical Churches
Western Christianity
Western Christian liturgical calendars are based on the cycle of the Roman Rite of the Catholic Church, including Lutheran, Anglican, and other Protestant calendars since this cycle pre-dates the Reformation. Generally, the liturgical seasons in western Christianity are Advent, Christmas, Ordinary Time (Time after Epiphany), Lent, Easter, and Ordinary Time (Time after Pentecost). Some Protestant traditions do not include Ordinary Time: every day falls into a denominated season.
Vanderbilt University Professor Hoyt L. Hickman, with regard to the calendars of the Western Christians Churches that use the Revised Common Lectionary, including Presbyterians, Methodists, Anglicans/Episcopalians, Lutherans and some Baptists, among others, states that:
Protestant Churches, with exception of the Lutheran and Anglican, generally observe fewer feasts with regard to the saints, than the aforementioned liturgical denominations, in addition to the Catholic and Orthodox Churches.
Denominational specifics
Catholic Church
The Catholic Church sets aside certain days and seasons of each year to recall and celebrate various events in the life of Christ. In its Roman Rite the liturgical year begins with Advent, the time of preparation for both the celebration of Jesus' birth, and his expected second coming at the end of time. This season lasts until Christmas Eve on December 24. Christmastide follows, beginning with First Vespers of Christmas on the evening of December 24 and ending with the Feast of the Baptism of the Lord. Traditionally, the end of Christmastide was February 2, or the Feast of the Presentation of the Lord, also known as Candlemas. This feast recounts the 40 days of rest Mary took before being purified and presenting her first-born son to the Temple in Jerusalem.
Lent is the period of purification and penance that begins on Ash Wednesday and ends on Holy Thursday. The Holy Thursday evening Mass of the Lord's Supper marks the beginning of the Easter Triduum, which includes Good Friday, Holy Saturday, and Easter Sunday. The days of the Easter Triduum recall Jesus' Last Supper with his disciples, death on the cross, burial, and resurrection. The seven-week liturgical season of Easter immediately follows the Triduum, climaxing at Pentecost. This last feast recalls the descent of the Holy Spirit upon Jesus' disciples after the Ascension of Jesus. The rest of the liturgical year is commonly known as Ordinary Time.
There are many forms of liturgy in the Catholic Church. Even putting aside the many Eastern rites in use, the Latin liturgical rites alone include the Ambrosian Rite, the Mozarabic Rite, and the Cistercian Rite, as well as other forms that have been largely abandoned in favour of adopting the Roman Rite.
Of this rite, what is now the "ordinary" or, to use a word employed in the Letter of Pope Benedict XVI accompanying the motu proprio Summorum Pontificum, the "normal" form is that given to it after the Second Vatican Council by Pope Paul VI and Pope John Paul II, while the 1962 Roman Missal form remains authorized, as an "extraordinary" form, for priests of the Latin Church without restriction in private celebrations, and under the conditions indicated in article 5 of the motu proprio Summorum Pontificum in public celebrations.
The liturgical calendar in that form of the Roman Rite (see General Roman Calendar) of 1960 differs in some respects from that of the present ordinary form, as will be noted below, and also from the earlier General Roman Calendar of Pope Pius XII, the still earlier General Roman Calendar of 1954 and the original Tridentine Calendar. These articles can be consulted with regard to the Roman-Rite liturgical year before 1960.
Lutheran Churches
Anglican Church
The Church of England, Mother Church of the Anglican Communion, uses a liturgical year that is in most respects identical to that of the 1969 Catholic Common Lectionary. While the calendars contained within the Book of Common Prayer and the Alternative Service Book (1980) have no "Ordinary Time", Common Worship (2000) adopted the ecumenical 1983 Revised Common Lectionary. The few exceptions are Sundays following Christmas and Transfiguration, observed on the last Sunday before Lent instead of on Reminiscere.
In some Anglican traditions (including the Church of England) the Christmas season is followed by an Epiphany season, which begins on the Eve of the Epiphany (on 6 January or the nearest Sunday) and ends on the Feast of the Presentation (on February 2 or the nearest Sunday). Ordinary Time then begins after this period.
The Book of Common Prayer contains within it the traditional Western Eucharistic lectionary which traces its roots to the Comes of St. Jerome in the 5th century. Its similarity to the ancient lectionary is particularly obvious during Trinity season (Sundays after the Sunday after Pentecost), reflecting that understanding of sanctification.
Reformed Churches
Reformed Christians emphasize weekly celebration of the Lord's day and, while some of them celebrate also what they call the five evangelical feasts, others celebrate no holy days.
Liturgical calendar
Advent
Advent (from the Latin word adventus, which means "arrival" or "coming") is the first season of the liturgical year. It begins four Sundays before Christmas, the Sunday falling on or nearest to November 30, and ends on Christmas Eve. Traditionally observed as a "fast", it focuses on preparation for the coming of Christ, not only the coming of the Christ-child at Christmas, but also, in the first weeks, on the eschatological final coming of Christ, making Advent "a period for devout and joyful expectation".
This season is often marked by the Advent Wreath, a garland of evergreens with four candles. Although the main symbolism of the advent wreath is simply marking the progression of time, many churches attach themes to each candle, most often 'hope', 'faith', 'joy', and 'love'. Other popular devotions during Advent include the use of the Advent Calendar or the Tree of Jesse to count down the days to Christmas.
Liturgical colour: violet or purple; blue in some traditions, such as Anglican/Episcopalian, Methodist, and Lutheran.
Christmastide
The Christmas season immediately follows Advent. The traditional Twelve Days of Christmas begin with Christmas Eve on the evening of December 24 and continue until the feast of Epiphany. The actual Christmas season continues until the Feast of the Baptism of Christ, which in the present form of the Roman Rite is celebrated on the Sunday after January 6, or the following Monday if that Sunday is Epiphany.
In the pre-1970 form, this feast is celebrated on January 13, unless January 13 is a Sunday, in which case the feast of the Holy Family is celebrated instead. Until the suppression of the Octave of the Epiphany in the 1960 reforms, January 13 was the Octave day of the Epiphany, providing the date for the end of the season.
Traditionally, the end of Christmastide was February 2, or the Feast of the Presentation of the Lord, also known as Candlemas. This feast recounts the 40 days of rest Mary took before being purified and presenting her first-born son to the Temple in Jerusalem. In medieval times, Candlemas eve (Feb. 1st) marked the day when all Christmas decorations, including the Christmas tree and the Nativity scene, were taken down. However, the tradition of ending Christmastide on Candlemas has slowly waned, except in some pockets of the Hispanic world where Candlemas (or La Fiesta de la Candelaria) is still an important feast and the unofficial end of the Christmas season.
Liturgical colour: white
Ordinary Time
"Ordinary" comes from the same root as our word "ordinal", and in this sense means "the counted weeks". In the Catholic Church and in some Protestant traditions, these are the common weeks which do not belong to a proper season. In Latin, these seasons are called the weeks , or "through the year".
In the current form of the Roman Rite adopted following the Second Vatican Council, Ordinary Time consists of 33 or 34 Sundays and is divided into two sections. The first portion extends from the day following the Feast of the Baptism of Christ until the day before Ash Wednesday (the beginning of Lent). It contains anywhere from three to eight Sundays, depending on how early or late Easter falls.
The main focus in the readings of the Mass is Christ's earthly ministry, rather than any one particular event. The counting of the Sundays resumes following Eastertide; however, two Sundays are replaced by Pentecost and Trinity Sunday, and depending on whether the year has 52 or 53 weeks, one may be omitted.
In the pre-1970 form of the Roman Rite, the Time after Epiphany has anywhere from one to six Sundays. As in the current form of the rite, the season mainly concerns Christ's preaching and ministry, with many of his parables read as the Gospel readings. The season begins on January 14 and ends on the Saturday before Septuagesima Sunday. Omitted Sundays after Epiphany are transferred to Time after Pentecost and celebrated between the Twenty-Third and the Last Sunday after Pentecost according to an order indicated in the Code of Rubrics, 18, with complete omission of any for which there is no Sunday available in the current year. Before the 1960 revisions, the omitted Sunday would be celebrated on the Saturday before Septuagesima Sunday, or, in the case of the Twenty-Third Sunday after Pentecost, on the Saturday before the Last Sunday after Pentecost.
Liturgical colour: green
Septuagesima/Pre-Lenten Season
Septuagesima (from the Latin word for "seventieth") is a two-and-a-half-week period before Lent. This pre-Lent season is present in the pre-1970 form of the Roman Rite and in some Protestant calendars. It is a transition from the first part of the season per annum to the season of Lent, and a preparation for the fasting and penance which begin on Ash Wednesday. Although most of the Divine Office remains the same as during the season per annum, certain customs of Lent are adopted, including the suppression of the "Alleluia", the replacement of the Alleluia at Mass with the Tract and the Gloria is no longer said on Sundays.
In the 1969 reform of the Roman Rite, this intermediate season was removed, with these weeks becoming part of Ordinary Time.
Liturgical colour (where observed): violet or purple
Lent and Passiontide
Lent is a major penitential season of preparation for Easter. It begins on Ash Wednesday and, if the penitential days of Good Friday and Holy Saturday are included, lasts for forty days, since the six Sundays within the season are not counted.
In the Roman Rite, the Gloria in Excelsis Deo and the Te Deum are not used in the Mass and Liturgy of the Hours respectively, except on Solemnities and Feasts, and the Alleluia and verse that usually precede the reading of the Gospel is either omitted or replaced with another acclamation.
Lutheran churches make these same omissions.
As in Advent, the deacon and subdeacon of the pre-1970 form of the Roman Rite do not wear their habitual dalmatic and tunicle (signs of joy) in Masses of the season during Lent; instead they wear "folded chasubles", in accordance with the ancient custom.
In the pre-1970 form of the Roman Rite, the two weeks before Easter form the season of Passiontide, a subsection of the Lenten season that begins with Matins of Ash Wednesday and ends immediately before the Mass of the Easter Vigil. In this form, what used to be officially called Passion Sunday, has the official name of the First Sunday in Passiontide, and Palm Sunday has the additional name of the Second Sunday in Passiontide. In Sunday and ferial Masses (but not on feasts celebrated in the first of these two weeks) the Gloria Patri is omitted at the Entrance Antiphon and at the Lavabo, as well as in the responds in the Divine Office.
In the post-1969 form of the Roman Rite, "Passion Sunday" and "Palm Sunday" are both names for the Sunday before Easter, officially called "Palm Sunday of the Lord's Passion". The former Passion Sunday became a fifth Sunday of Lent. The earlier form reads Matthew's account on Sunday, Mark's on Tuesday, and Luke's on Wednesday, while the post-1969 form reads the Passion only on Palm Sunday (with the three Synoptic Gospels arranged in a three-year cycle) and on Good Friday, when it reads the Passion according to John, as also do earlier forms of the Roman Rite.
The veiling of crucifixes and images of the saints with violet cloth, which was obligatory before 1970, is left to the decision of the national bishops' conferences. In the United States, it is permitted but not required, at the discretion of the pastor. In all forms, the readings concern the events leading up to the Last Supper and the betrayal, Passion, and death of Christ.
The week before Easter is called Holy Week.
In the Roman Rite, feasts that fall within that week are simply omitted, unless they have the rank of Solemnity, in which case they are transferred to another date. The only solemnities inscribed in the General Calendar that can fall within that week are those of St. Joseph and the Annunciation.
Liturgical colour: violet or purple. The colour rose may be used, where it is the practice, on Laetare Sunday (4th Sunday of Lent). On Palm Sunday the colour since 1970 is red, by earlier rules violet or purple, with red being used after 1955 for the blessing of the palms.
Easter Triduum
The Easter Triduum consists of Good Friday, Holy Saturday and Easter Sunday. Each of these days begins liturgically not with the morning but with the preceding evening.
The triduum begins on the evening before Good Friday with Mass of the Lord's Supper, celebrated with white vestments, and often includes a ritual of ceremonial footwashing. It is customary on this night for a vigil involving private prayer to take place, beginning after the evening service and continuing until midnight. This vigil is occasionally renewed at dawn, continuing until the Good Friday liturgy.
During the day of Good Friday Mass is not celebrated in the Catholic Church. Instead a Celebration of the Passion of the Lord is held in the afternoon or evening. It consists of three parts: a Liturgy of the Word that includes the reading of the account of the Passion by John the Evangelist and concludes with a solemn Universal Prayer. Other churches also have their Good Friday commemoration of the Passion.
The colour of vestments varies: no colour, red, or black are used in different traditions. Coloured hangings may be removed. Lutheran churches often either remove colourful adornments and icons, or veil them with drab cloth. The service is usually plain with somber music, ending with the congregation leaving in silence. In the Catholic, some Lutheran, and High Anglican rites, a crucifix (not necessarily the one which stands on or near the altar on other days of the year) is ceremoniously unveiled. Other crucifixes are unveiled, without ceremony, after the service.
Holy Saturday commemorates the day during which Christ lay in the tomb. In the Catholic Church, there is no Mass on this day; the Easter Vigil Mass, which, though celebrated properly at the following midnight, is often celebrated in the evening, is an Easter Mass. With no liturgical celebration, there is no question of a liturgical colour.
The Easter Vigil is held in the night between Holy Saturday and Easter Sunday, to celebrate the resurrection of Jesus. See also Paschal candle. The liturgical colour is white, often together with gold. In the Roman Rite, during the "Gloria in Excelsis Deo" the organ and bells are used in the liturgy for the first time in 2 days, and the statues, which have been veiled during Passiontide (at least in the Roman Rite through the 1962 version), are unveiled. In Lutheran churches, colours and icons are re-displayed as well.
Eastertide
Easter is the celebration of Jesus' Resurrection. The date of Easter varies from year to year, according to a lunar-calendar dating system (see computus for details). In the Roman Rite, the Easter season extends from the Easter Vigil through Pentecost Sunday. In the pre-1970 form of the rite, this season includes also the Octave of Pentecost, so Eastertide lasts until None of the following Saturday.
In the Roman Rite, the Easter octave allows no other feasts to be celebrated or commemorated during it; a solemnity, such as the Annunciation, falling within it is transferred to the following Monday. If Easter Sunday or Easter Monday falls on 25 April, the Greater Litanies, which in the pre-1970 form of the Roman Rite are on that day, are transferred to the following Tuesday.
By a decree of May 5, 2000, the Second Sunday of Easter (the Sunday after Easter Day itself), is known also in the Roman Rite as the Feast of the Divine Mercy.
Ascension Thursday, which celebrates the return of Jesus to heaven following his resurrection, is the fortieth day of Easter, but, in places where it is not observed as a Holy Day of Obligation, the post-1969 form of the Roman rite transfers it to the following Sunday.
Pentecost is the fiftieth and last day of the Easter season. It celebrates the sending of the Holy Spirit to the Apostles, which traditionally marks the birth of the Church, see also Apostolic Age.
Liturgical colour: white, but red on the feast of Pentecost.
Ordinary Time, Time after Pentecost, Time after Trinity, or Kingdomtide
This season, under various names, follows the Easter season and the feasts of Easter, Ascension, and Pentecost. In the post-1969 form of the Roman rite, Ordinary Time resumes on Pentecost Monday, omitting the Sunday which would have fallen on Pentecost. In the earlier form, where Pentecost is celebrated with an octave, the Time after Pentecost begins at Vespers on the Saturday after Pentecost. The Sundays resume their numbering at the point that will make the Sunday before Advent the thirty-fourth, omitting any weeks for which there is no room (present-day form of the Roman Rite) or are numbered as "Sundays after Pentecost" (pre-1970 Roman Rite, Eastern Orthodoxy and some Protestants) or as "Sundays after Trinity" (some Protestants). This season ends on the Saturday before the First Sunday of Advent.
Feasts during this season include:
Trinity Sunday, the first Sunday after Pentecost
Solemnity of the Most Holy Body and Blood of Christ (Roman Rite and some Anglican and Lutheran traditions), Thursday of the second week after Pentecost, often celebrated on the following Sunday
Solemnity of the Most Sacred Heart of Jesus (Roman Rite), Friday of the third week after Pentecost
Assumption of Mary on 15 August
Feast of Christ the King, last Sunday before Advent (Roman Rite, Lutherans, Anglicans) or last Sunday in October (1925–1969 form of the Roman Rite)
In the final few weeks of Ordinary Time, many churches direct attention to the coming of the Kingdom of God, thus ending the liturgical year with an eschatological theme that is one of the predominant themes of the season of Advent that began the liturgical year. For instance, in the extraordinary form of the Roman Rite, the Gospel of the Last Sunday is and in the ordinary form of the Roman Rite all the last three Sundays of the liturgical year are affected by the theme of the Second Coming.
While the Roman Rite adopts no special designation for this final part of Ordinary Time, some denominations do, and may also change the liturgical colour. The Church of England uses the term "Sundays before Advent" for the final four Sundays and permits red vestments as an alternative. The United Methodist Church may use the name "Kingdomtide". The Lutheran Church–Missouri Synod (LCMS) uses the terms "Third-Last, Second-Last and Last Sunday in the Church Year" and does not change from green. The LCMS does not officially celebrate a "Feast of Christ the King." The Wisconsin Evangelical Lutheran Synod (WELS) uses the term "Period of End Times" and assigns red vestments to the first and second Sundays.
Calendar of saints
In some Protestant traditions, especially those with closer ties to the Lutheran tradition, Reformation Sunday is celebrated on the Sunday preceding October 31, commemorating the purported day Martin Luther posted the 95 Theses on the door of the Castle Church in Wittenberg. The liturgical colour is red, celebrating the Holy Spirit's continuing work in renewing the Church.
Most Western traditions celebrate All Saints' Day (All Hallow's Day) on November 1 or the Sunday following, with the eve of this feast, All Hallow's Eve being October 31. The liturgical colour is white. The following day, November 2, is All Souls' Day. The period including these days is often referred to as Allhallowtide or Allsaintstide.
Saints Days are observed by Lutherans and include the apostles, Virgin Mary and noteworthy figures in the Christian faith. The Confession of St. Peter Week of Prayer for Christian Unity starting on January 18. Conversion of St. Paul ended week of prayer on January 25. Martin Luther King Jr., renewer of society, martyr January 15 (Evangelical Lutheran Church in America only), Presentation of Our Lord and Purification of the Mary Candlemas on February 2. Joseph, Guardian of Jesus St Joseph on March 19, Annunciation March 25, Visitation of Mary on May 31.
Lutherans also celebrate St John the Baptist or the Beheading of St John the Baptist on June 24, St Mary Magdalene July 22, St. Mary, Mother of Our Lord or the Assumption of the Blessed Virgin Mary on August 15, Holy Cross Day September 14, Francis of Assisi, renewal of the Church St. Francis of Assisi on October 4, and the Holy Innocents, Martyrs December 28.
Lesser Feasts and Commemorations on the Lutheran liturgical calendar include Anthony of Egypt on January 17, Henry, Bishop of Uppsala, martyr Henry of Uppsala on January 19, Timothy, Titus and Silas, missionaries St Timothy, St Titus and St Silas Day on January 26, Ansgar, Bishop of Hamburg, missionary to Denmark and Sweden St Ansgar on February 3, Cyril, monk and Methodius, bishop, missionaries to the Slavs St Cyril and St Methodius on February 14, Gregory the Great on March 12, St Patrick on March 17, Olavus Petri, priest and Laurentius Petri, Bishop of Uppsala, on April 19, St Anselm on April 21, Catherine of Siena on April 29, St Athanasius on May 2, St Monica on May 4, Eric IX of Sweden on May 18, St Boniface on June 5, Basil the Great, Gregory of Nyssa and Gregory of Nazianzus on June 14, Benedict of Nursia on July 11, Birgitta of Sweden on July 23, St Anne, Mother of Mary on July 26, St Dominic on August 8, Augustine of Hippo on August 28, St Cyprian on September 16, Teresa of Avila on October 15, Martin de Porres on November 3, Martin of Tours on November 11, Elizabeth of Hungary on November 17, St Lucy on December 13. There are many other holy days in the Lutheran calendar.
Some traditions celebrate St. Michael's Day (Michaelmas) on September 29.
Some traditions celebrate St. Martin's Day (Martinmas) on November 11.
Liturgical colours: white if the saint was not martyred; red if the saint was martyred
Hierarchy of feast days
There are degrees of solemnity of the office of the feast days of saints. In the 13th century, the Roman Rite distinguished three ranks: simple, semidouble and double, with consequent differences in the recitation of the Divine Office or Breviary. The simple feast commenced with the chapter (capitulum) of First Vespers, and ended with None. It had three lessons and took the psalms of Matins from the ferial office; the rest of the office was like the semidouble. The semidouble feast had two Vespers, nine lessons in Matins, and ended with Compline. The antiphons before the psalms were only intoned. In the Mass, the semidouble had always at least three "orationes" or collects. On a double feast the antiphons were sung in their entirety, before and after the psalms, while in Lauds and Vespers there were no suffragia of the saints, and the Mass had only one "oratio" (if no commemoration was prescribed). If ordinary double feasts (referred to also as lesser doubles) occurred with feasts of a higher rank, they could be simplified, except the octave days of some feasts and the feasts of the Doctors of the Church, which were transferred.
To the existing distinction between major and ordinary or minor doubles, Pope Clement VIII added two more ranks, those of first-class or second-class doubles. Some of these two classes were kept with octaves. This was still the situation when the 1907 article Ecclesiastical Feasts in the Catholic Encyclopedia was written. In accordance with the rules then in force, feast days of any form of double, if impeded by "occurrence" (falling on the same day) with a feast day of higher class, were transferred to another day.
Pope Pius X simplified matters considerably in his 1911 reform of the Roman Breviary. In the case of occurrence the lower-ranking feast day could become a commemoration within the celebration of the higher-ranking one. Until then, ordinary doubles took precedence over most of the semidouble Sundays, resulting in many of the Sunday Masses rarely being said. While retaining the semidouble rite for Sundays, Pius X's reform permitted only the most important feast days to be celebrated on Sunday, although commemorations were still made until Pope John XXIII's reform of 1960.
The division into doubles (of various kinds) semidoubles and simples continued until 1955, when Pope Pius XII abolished the rank of semidouble, making all the previous semidoubles simples, and reducing the previous simples to a mere commemoration in the Mass of another feast day or of the feria on which they fell (see General Roman Calendar of Pope Pius XII).
Then, in 1960, Pope John XXIII issued the Code of Rubrics, completely ending the ranking of feast days by doubles etc., and replacing it by a ranking, applied not only to feast days but to all liturgical days, as I, II, III, and IV class days.
The 1969 revision by Pope Paul VI divided feast days into "solemnities", "feasts" and "memorials", corresponding approximately to Pope John XXIII's I, II and III class feast days. Commemorations were abolished. While some of the memorials are considered obligatory, others are optional, permitting a choice on some days between two or three memorials, or between one or more memorials and the celebration of the feria. On a day to which no obligatory celebration is assigned, the Mass may be of any saint mentioned in the Roman Martyrology for that day.
Assumption of Mary
Observed by Catholics and some Anglicans on August 15, which is the same as the Eastern and Orthodox feast of the Dormition, the end of the earthly life of the Virgin Mary and, for some, her bodily Assumption into heaven, is celebrated. The Catholic teaching on this feast was defined as dogma on November 1, 1950, by Pope Pius XII in the Papal Bull, Munificentissimus Deus.
In other Anglican and Lutheran traditions, as well as a few others, August 15 is celebrated as St. Mary, Mother of the Lord.
Liturgical colour: white
Secular observance
Because of the dominance of Christianity in Europe throughout the Middle Ages, many features of the Christian year became incorporated into the secular calendar. Many of its feasts (e.g., Christmas, Mardi Gras, Saint Patrick's Day) remain holidays, and are now celebrated by people of all faiths and none — in some cases worldwide. The secular celebrations bear varying degrees of likeness to the religious feasts from which they derived, often also including elements of ritual from pagan festivals of similar date.
See also
Christian worship
Quarter tense
References
Further reading
Stookey, L. H. Calendar: Christ's Time for the Church, 1996.
Hickman, Hoyt L., et al. Handbook of the Christian Year, 1986.
Webber, Robert E. Ancient-Future Time: Forming Spirituality through the Christian Year, 2004.
Schmemann, Fr. Alexander. The Church Year (Celebration of Faith Series, Sermons Vol. 2), 1994.
Talley, Thomas J. The Origins of the Liturgical Year, Ed. 2. 1991.
External links
The Catholic Church's liturgical calendar, from US Catholic Bishops , or from O.S.V. publishing .
Universalis — A liturgical calendar of the Catholic Church including the Liturgy of the Hours and the Mass readings.
Greek Orthodox Calendar – Greek Orthodox Calendar & Online Chapel
Russian Orthodox Calendar at Holy Trinity Russian Orthodox Church
Lectionary Central – For the study and use of the traditional Western Eucharistic lectionary (Anglican).
Seasons
Christian terminology
Types of year |
null | null | Tennessee Valley Authority | eng_Latn | The Tennessee Valley Authority (TVA) is a federally-owned electric utility corporation in the United States. TVA's service area covers all of Tennessee, portions of Alabama, Mississippi, and Kentucky, and small areas of Georgia, North Carolina, and Virginia. While owned by the federal government, TVA receives no taxpayer funding and operates similar to a private for-profit company. It is headquartered in Knoxville, Tennessee, and is the sixth largest power supplier and largest public utility in the country.
The TVA was created by Congress in 1933 as part of President Franklin D. Roosevelt's New Deal. Its initial purpose was to provide navigation, flood control, electricity generation, fertilizer manufacturing, regional planning, and economic development to the Tennessee Valley, a region which was significantly affected by the Great Depression relative to the rest of the nation. TVA was envisioned both as a power supplier and a regional economic development agency that would work to help modernize the region's economy and society, but later evolved primarily into an electric utility. It was the first large regional planning agency of the U.S. federal government and remains the largest.
Under the leadership of David E. Lilienthal, the TVA also became the global model for the United States' later efforts to help modernize agrarian societies in the developing world. Historically, the TVA has been documented as a success in its efforts to modernize the Tennessee Valley and helping to recruit new employment opportunities to the region. Despite its successes, historians have criticized its use of eminent domain with the displacement of over 125,000 Tennessee Valley residents for the agency's infrastructure projects.
Operation
The Tennessee Valley Authority was initially founded as an agency to provide general economic development to the region through power generation, flood control, navigation assistance, fertilizer manufacturing, and agricultural development. Since the Depression years, it has developed primarily into a power utility. Despite its shares being owned by the federal government, TVA operates like a private corporation, and receives no taxpayer funding. The TVA Act authorizes the company to use eminent domain.
TVA provides electricity to approximately ten million people through a diverse portfolio that includes nuclear, coal-fired, natural gas-fired, hydroelectric, and renewable generation. TVA sells its power to 154 local power utilities, 5 direct industrial and institutional customers, and 12 area utilities. In addition to power generation, TVA provides flood control with its 29 hydroelectric dams. Resulting lakes and other areas also allow for recreational activities. The TVA also provides navigation and land management along rivers within its region of operation. TVA also assists governments and private companies on economic development projects.
TVA's headquarters are located in downtown Knoxville, with large administrative offices in Chattanooga (training/development; supplier relations; power generation and transmission) and Nashville (economic development) in Tennessee and Muscle Shoals, Alabama. TVA was originally headquartered in Muscle Shoals, but gradually moved its headquarters to Knoxville. At one point, TVA's headquarters were housed in the Old Federal Customs House at the corner of Clinch Avenue and Market Street. The building is now operated as a museum and is listed on the National Register of Historic Places.
The Tennessee Valley Authority Police are the primary law enforcement agency for the company. Initially part of the TVA, in 1994 the TVA Police was authorized as a federal law enforcement agency.
Board of Directors
The Tennessee Valley Authority is governed by a nine member part-time Board of Directors, nominated by the President of the United States, and confirmed by the Senate. A minimum of seven of the directors are required to be a resident of TVA's service area. The members select the Chairman from their number, and serve five-year terms. They receive an annual stipend of $45,000 and $50,000 for the chairman. The board members choose the TVA's chief executive officer (CEO).
History
Background
During the 1920s and the 1930s, Americans began to support the idea of public ownership of utilities, particularly hydroelectric power facilities. Many believed privately owned power companies were charging too much for power, did not employ fair operating practices, and were subject to abuse by their owners (utility holding companies), at the expense of consumers.
The concept of government-owned generation facilities selling to publicly owned distribution utilities was controversial, however, and remains so today.
During his 1932 presidential campaign, Franklin D. Roosevelt expressed his belief that private utilities had "selfish purposes" and said, "Never shall the federal government part with its sovereignty or with its control of its power resources while I'm president of the United States." Senator George W. Norris of Nebraska also distrusted private utility companies, and in 1920 blocked a proposal from industrialist Henry Ford to build a private dam and create a utility to modernize the Tennessee Valley. The private sector practice of forming utility holding companies had resulted in them controlling 94 percent of generation by 1921, and they were essentially unregulated. In an effort to change this, Congress and Roosevelt enacted the Public Utility Holding Company Act of 1935 (PUHCA).
In 1930, Norris sponsored the Muscle Shoals Bill, which would have built a federal dam in the valley, but it was vetoed by President Herbert Hoover, who believed it to be socialistic. The idea behind the Muscle Shoals project became a core part of President Franklin D. Roosevelt's New Deal program that created the Tennessee Valley Authority, however.
Even by Depression standards, the Tennessee Valley was in dire economic straits in 1933. Thirty percent of the population was affected by malaria. The average income in the rural areas was $639 per year (equivalent to $ in ), with some families surviving on as little as $100 per year (equivalent to $ in ). Much of the land had been exhausted by poor farming practices, and the soil was eroded and depleted Crop yields had fallen, reducing farm incomes. The best timber had been cut, and 10% of forests were lost to fires each year.
Early history
President Franklin Delano Roosevelt signed the Tennessee Valley Authority Act (ch. 32, , codified as amended at , et seq.), creating the TVA. The agency was initially tasked with modernizing the region, using experts and electricity to combat human and economic problems. TVA developed fertilizers, and taught farmers ways to improve crop yields. In addition, it helped replant forests, control forest fires, and improve habitat for fish and wildlife.
The Authority hired many of the area's unemployed for a variety of jobs: they conducted conservation, economic development, and social programs. For instance, a library service was instituted for this area. The professional staff at headquarters were generally composed of experts from outside the region. By 1934, TVA employed more than 9,000 people. The workers were classified by the usual racial and gender lines of the region, which limited opportunities for minorities and women. TVA hired a few African Americans, generally restricted for janitorial or other low-level positions. TVA recognized labor unions; its skilled and semi-skilled blue collar employees were unionized, a breakthrough in an area known for corporations hostile to miners' and textile workers' unions. Women were excluded from construction work.
Many local landowners were suspicious of government agencies, but TVA successfully introduced new agricultural methods into traditional farming communities by blending in and finding local champions. Tennessee farmers would often reject advice from TVA officials, so the officials had to find leaders in the communities and convince them that crop rotation and the judicious application of fertilizers could restore soil fertility. Once they had convinced the leaders, the rest followed.
TVA immediately embarked on the construction of several hydroelectric dams, with the first, Norris Dam in upper East Tennessee, breaking ground on October 1, 1933. These facilities, designed with the intent of also controlling floods, greatly improved the lives of farmers and rural residents, making their lives easier and farms in the Tennessee Valley more productive. They also provided new employment opportunities to the poverty-stricken regions in the Valley. At the same time, however, they required the displacement of more than 125,000 valley residents or roughly 15,000 families, as well as some cemeteries and small towns, which caused some to oppose the projects, especially in rural areas. The projects also inundated several Native American archaeological sites, and graves were reinterred at new locations, along with new tombstones.
The available electricity attracted new industries to the region, including textile mills, providing desperately needed jobs, many of which were filled by women. A few regions of the Tennessee Valley did not receive electricity until the late 1940s and early 1950s, however. TVA was one of the first federal hydropower agencies, and was quickly hailed as a success. While most of the nation's major hydropower systems are federally managed today, other attempts to create similar regional corporate agencies have failed. The most notable was the proposed Columbia Valley Authority for the Columbia River in the Pacific Northwest, which was modeled off of TVA, but did not gain approval.
World War II
During World War II, the U.S. needed greater aluminum supplies to build airplanes. Aluminum plants required large amounts of electricity. To provide the power, TVA engaged in one of the largest hydropower construction programs ever undertaken in the U.S. By early 1942, when the effort reached its peak, 12 hydroelectric plants and one coal-fired steam plant were under construction at the same time, and design and construction employment reached a total of 28,000. In its first eleven years, TVA constructed a total of 16 hydroelectric dams.
The largest project of this period was the Fontana Dam. After negotiations led by then-Vice-President Harry Truman, TVA purchased the land from Nantahala Power and Light, a wholly owned subsidiary of Alcoa, and built Fontana Dam. Also in 1942, TVA's first coal-fired plant, the 267-megawatt Watts Bar Steam Plant, began operation. The government originally intended the electricity generated from Fontana to be used by Alcoa factories. However, the abundance of TVA power was one of the major factors in the decision by the U.S. Army to locate uranium enrichment facilities in Oak Ridge, Tennessee for the world's first atomic bombs. This was part of an effort codenamed the Manhattan Project.
Increasing power demand
By the end of World War II, TVA had completed a navigation channel the length of the Tennessee River and had become the nation's largest electricity supplier. Even so, the demand for electricity was outstripping TVA's capacity to produce power from hydroelectric dams, and so TVA began to construct additional coal-fired plants. Political interference kept TVA from securing additional federal appropriations to do so, so it sought the authority to issue bonds. Several of TVA's coal-fired plants, including Johnsonville, Widows Creek, Shawnee, Kingston, Gallatin, and John Sevier, began operations in the 1950s. In 1955 coal surpassed hydroelectricity as TVA's top generating source. On August 6, 1959, President Dwight D. Eisenhower signed into law an amendment to the TVA act, making the agency self-financing. During the 1950s, TVA's generating capacity nearly quadrupled.
The 1960s were years of further unprecedented economic growth in the Tennessee Valley. Capacity growth during this time slowed, but ultimately increased 56% between 1960 and 1970. To handle a projected future increase in electrical consumption, TVA began constructing 500 kilovolt (kV) transmission lines, the first of which was placed into service on May 15, 1965. Electric rates were among the nation's lowest during this time and stayed low as TVA brought larger, more efficient generating units into service. Plants completed during this time included Paradise, Bull Run, and Nickajack Dam. Expecting the Valley's electric power needs to continue to grow, TVA began building nuclear power plants in 1966 as a new source of power. During the 1960s and 1970s, TVA was engaged in what was up to that time its most controversial project – the Tellico Dam Project. The project was initially conceived in the 1940s but not completed until 1979.
1970s and 1980s
Significant changes occurred in the economy of the Tennessee Valley and the nation, prompted by energy crises in 1973 and 1979 and accelerating fuel costs throughout the decade. The average cost of electricity in the Tennessee Valley increased fivefold from the early 1970s to the early 1980s. TVA's first nuclear reactor, Browns Ferry Unit 1, began commercial operation on August 1, 1974. In the early 1970s, TVA set out to construct a total of 17 nuclear reactors, due to a projection of further rapid increase in power demand. However, in the 1980s, ten of these reactors were cancelled. On August 6, 1981, the Tennessee Valley Authority Board voted to defer the Phipps Bend plant, as well as to slow down construction on all other projects. The Hartsville and Yellow Creek plants were cancelled in 1984 and Bellefonte in 1988.
Construction of the Tellico Dam became controversial for political and environmental reasons, as laws had changed since early development in the valley. Scientists and other researchers had become more aware of the massive environmental effects of the dams and new lakes, and worried about preserving habitats and species. The Tellico Dam project was initially delayed because of concern over the snail darter, a threatened species. A lawsuit was filed under the Endangered Species Act and the U.S. Supreme Court ruled in favor of protecting the snail darter in Tennessee Valley Authority v. Hill in 1978. The project would also receive controversy regarding the intent of the dam, as the project's main motive was to support recreational and tourism development, unlike earlier dams constructed by TVA. Land acquired by eminent domain for the Tellico Dam and its reservoir, specifically land that encountered minimal inundation, was sold to private developers for the construction of present-day Tellico Village, a planned retirement community.
The cancellation of several of the planned nuclear plants rendered the agency in deep financial trouble. Marvin T. Runyon became chairman of the TVA in January 1988. During his four-year tenure he claimed to reduce management layers, cut overhead costs by more than 30%, and achieved cumulative savings and efficiency improvements of $1.8 billion. He also claimed to have revitalized the nuclear program and instituted a rate freeze that continued for ten years.
1990s to late 2010s
As the electric-utility industry moved toward restructuring and deregulation, TVA began preparing for competition. It cut operating costs by nearly $950 million a year, reduced its workforce by more than half, increased the generating capacity of its plants, and developed a plan to meet the energy needs of the Tennessee Valley through the year 2020.
In 1996, Watts Bar Unit 1 began operation. This was the last commercial nuclear reactor in the United States to begin operation in the 20th century. In 2002, TVA began work to restart a previously mothballed nuclear reactor at Browns Ferry Unit 1, which was completed in May 2007. In 2004, TVA implemented recommendations from the Reservoir Operations Study (ROS) on how it operates the Tennessee River system (the nation's fifth largest). In 2005, the TVA announced its intention to construct an Advanced Pressurized Water Reactor at its Bellefonte site in Alabama, filing the necessary applications in November 2007. In 2007 it announced plans to complete the unfinished Unit 2 at Watts Bar, which began commercial operation in October 2016. Watts Bar Unit 2 is the first, and so far only, new nuclear reactor to enter service in the United States in the 21st century.
On December 22, 2008, an earthen dike at TVA's Kingston Fossil Plant broke, spreading one billion gallons of wet coal ash across of land and into the tributaries of the Tennessee River. This produced damage from high levels of metal in the river. The TVA Office of the Inspector General's report, Inspection 2008-12283-02, Review of the Kingston Fossil Plant Ash Spill Cause Study and Observations About Ash Management, concluded that TVA culture had contributed to the spill.
In 2009, to gain more access to sustainable, green energy, TVA signed 20-year power purchase agreements with Maryland-based CVP Renewable Energy Co. and Chicago-based Invenergy Wind LLC for electricity generated by wind farms. In April 2011, TVA reached an agreement with the Environmental Protection Agency (EPA), four state governments, and three environmental groups to drastically reduce pollution and carbon emissions. Under the terms of the agreement, TVA was required to retire at least 18 of its 59 coal-fired units by the end of 2018, and install scrubbers in several others or convert them to make them cleaner, at a cost of $25 billion, by 2021. As a result, TVA closed several of its coal-fired power plants in the 2010s, converting some to natural gas. These include John Sevier in 2012, Shawnee Unit 10 in 2014, Widows Creek in 2015, Colbert in 2016, Johnsonville and Paradise Units 1 and 2 in 2017, Allen in 2018, and Paradise Unit 3 in 2020.
Recent history
In 2018, TVA opened a new cybersecurity center in its downtown Chattanooga Office Complex. More than 20 Information Technology specialists monitor emails, Twitter feeds and network activity for cybersecurity threats and threats to grid security. Across TVA's digital platform, 2 billion activities occur each day. The center is staffed 24 hours a day to spot any threats to TVA's 16,000 miles of transmission lines.
Given continued economic pressure on the coal industry, the TVA board defied President Donald Trump and voted in February 2019 to close two aging coal plants, Paradise 3 and Bull Run. TVA chief executive Bill Johnson said the decision was not about coal, per se, but rather "about keeping rates as low as feasible." The TVA stated that decommissioning the two plants would reduce its carbon output by about 4.4% annually. TVA announced in April 2021 plans to completely phase out coal power by 2035. The following month, the TVA board voted to consider replacing almost all of their operating coal facilities with combined-cycle gas plants. Such plants considered for gas plant redevelopment include the Cumberland, Gallatin, Shawanee, and Kingston facilities.
In early February 2020, TVA awarded an outside company, Framatome, several multi-million-dollar contracts for work across the company's reactor fleet. This includes fuel for the Browns Ferry Nuclear Plant, fuel handling equipment upgrades across the fleet and steam generator replacements at the Watts Bar Nuclear Plant. Framatome will provide its state-of-the-art ATRIUM 11 fuel for the three boiling water reactors at Browns Ferry with the first use planned for 2023. This contract makes TVA the third U.S. utility to switch to the ATRIUM 11 fuel design. On August 3, 2020, president Donald Trump fired the TVA chairman and another board member, saying they were overpaid and had outsourced 200 high-tech jobs. The move came after U.S. Tech Workers, a nonprofit that works to limit visas given to foreign technology workers, criticized the TVA for laying off its own workers and replacing them with contractors using foreign workers with H-1B visas.
Citing its plan to reach net-zero carbon emissions in 2050, the TVA Board voted to approve an advanced approach of nuclear energy technology with an estimated $200 million investment, known as the New Nuclear Program (NNP) in February 2022. The NNP would promote the construction of new nuclear power facilities, particularly small modular reactors, with the first facility being constructed in partnership with Oak Ridge National Laboratory at the Clinch River Nuclear Site in Oak Ridge.
Facilities
Power stations
With a generating capacity of approximately 35 gigawatts (GW), TVA has the sixth highest generation capacity of any utility company in the United States. TVA's power mix as of 2022 is five coal-fired power plants, 29 hydroelectric dams, three nuclear plants (with seven operating reactors), nine simple-cycle natural gas combustion turbine plants, nine combined cycle gas plants, 1 pumped storage hydroelectric plant, 1 wind energy site, and 15 small solar energy sites. In fiscal year 2020, nuclear generation made up about 41% of TVA's total, natural gas 26%, coal 14%, hydroelectric 13%, and wind and solar 3%. TVA purchases about 15% of the power it sells from other power producers, which includes power from combined cycle natural gas plants, coal plants, and wind installations, and other renewables. The cost of Purchased Power is part of the "Fuel Cost Adjustment" (FCA) charge that is separate from the TVA Rate. Watts Bar Nuclear Plant produces tritium as a byproduct for the U.S. National Nuclear Security Administration, which requires tritium for nuclear weapons (for "boosted" fission primaries and for fusion secondaries).
Electric transmission
TVA owns and operates their own electric grid, which consists of approximately of lines, one of the largest grids in the United States. This grid is part of the Eastern Interconnection of the North American power transmission grid, and is under the jurisdiction of the SERC Reliability Corporation. Like most North American utilities, TVA uses a maximum transmission voltage of 500 kilovolts (kV), with lines carrying this voltage using bundled conductors with three conductors per phase. The vast majority of TVA's transmission lines carry 161 kV, with the company also operating a number of sub-transmission lines with a voltage of 69 kV.
Recreation
TVA has conveyed approximately of property for recreation and preservation purposes including public parks, public access areas and roadside parks, wildlife refuges, national parks and forests, and other camps and recreation areas, comprising approximately 759 different sites.
Megasites
To qualify for a TVA Megasite certificate the qualifications are at least 1,000 acres, with interstate access, the potential for rail service, environmental impact study, and utility service capable of serving a major manufacturing facility. Seven TVA Megasites have been developed so far with capital investments of over $5 billion.
Locations:
Huntsville
Chattanooga
Golden Triangle (Mississippi) (2 sites)
Hopkinsville, Kentucky
Memphis Regional Megasite
West Kentucky Megasite
Criticism and controversies
Allegations of federal government overreach
TVA was heralded by New Dealers and the New Deal Coalition not only as a successful economic development program for a depressed area but also as a democratic nation-building effort overseas because of its alleged grassroots inclusiveness as articulated by director David E. Lilienthal. However, the TVA was controversial early on, as some believed its creation was an overreach by the federal government.
Supporters of TVA note that the agency's management of the Tennessee River system without appropriated federal funding saves federal taxpayers millions of dollars annually. Opponents, such as Dean Russell in The TVA Idea, in addition to condemning the project as being socialistic, argued that TVA created a "hidden loss" by preventing the creation of "factories and jobs that would have come into existence if the government had allowed the taxpayers to spend their money as they wished." Defenders note that TVA is overwhelmingly popular in Tennessee among conservatives and liberals alike, as Barry Goldwater discovered in 1964, when he proposed selling the agency.
Historian Thomas McCraw concludes that Roosevelt "rescued the [power] industry from its own abuses" but "he might have done this much with a great deal less agitation and ill will". New Dealers hoped to build numerous other federal utility corporations around the country but were defeated by Wendell Willkie and the Conservative coalition in Congress. The valley authority model did not replace the limited-purpose water programs of the Bureau of Reclamation and the Army Corps of Engineers. State-centered theorists hold that reformers are most likely to succeed during periods such as the New Deal era, when they are supported by a democratized polity and when they dominate Congress and the administration.
However, it has been shown that in river policy, the strength of opposing interest groups also mattered. The TVA bill was passed in 1933 because reformers like Norris skillfully coordinated action at potential choke points and weakened the already disorganized opponents among the electric power industry lobbyists. In 1936, however, after regrouping, opposing river lobbyists and conservative coalition congressmen took advantage of the New Dealers' spending mood by expanding the Army Corps' flood control program. They also helped defeat further valley authorities, the most promising of the New Deal water policy reforms.
When Democrats after 1945 proclaimed the Tennessee Valley Authority as a model for countries in the developing world to follow, conservative critics charged it was a top-heavy, centralized, technocratic venture that displaced locals and did so in insensitive ways. Thus, when the program was used as the basis for modernization programs in various parts of the third world during the Cold War, such as in the Mekong Delta in Vietnam, its failure brought a backlash of cynicism toward modernization programs that has persisted.
Then-movie star Ronald Reagan had moved to television as the host and a frequent performer for General Electric Theater during 1954. Reagan was later fired by General Electric in 1962 in response to his publicly referring to the TVA (TVA being a major customer for GE turbines) as one of the problems of "big government". Some claim that Reagan was instead fired due to a criminal antitrust investigation involving him and the Screen Actors Guild. However, Reagan was only interviewed; nobody was actually charged with anything in the investigation.
In 1963, U.S. Senator and Republican presidential candidate Barry Goldwater was quoted in a Saturday Evening Post article by Stewart Alsop as saying, "You know, I think we ought to sell TVA." He had called for the sale to private companies of particular parts of the Authority, including its fertilizer production and steam-generation facilities, because "it would be better operated and would be of more benefit for more people if it were part of private industry." Goldwater's quotation was used against him in a TV ad by Doyle Dane Bernbach for President Lyndon Johnson's 1964 campaign, which depicted an auction taking place atop a dam. It was voiced over as follows: "In a Saturday Evening Post article dated August 31, 1963, Barry Goldwater said, 'You know, I think we ought to sell TVA.' This is a promise: President Johnson will not sell TVA. Vote for him on November 3. The stakes are too high for you to stay home."
Legal challenges
TVA faced multiple constitutional challenges. The United States Supreme Court ruled TVA to be constitutional in Ashwander v. Tennessee Valley Authority (297 U.S. 288) in 1936. The Court noted that regulating commerce among the states includes regulation of streams and that controlling floods is required for keeping streams navigable; it also upheld the constitutionality of the TVA under the War Powers Clause, seeing its activities as a means of assuring the electric supply for the manufacture of munitions in the event of war. The argument before the court was that electricity generation was a by-product of navigation and flood control and therefore could be considered constitutional. The CEO of the Tennessee Electric Power Company (TEPCO), Jo Conn Guild, was vehemently opposed to the creation of TVA, and with the help of attorney Wendell Willkie, challenged the constitutionality of the TVA Act in federal court. The U.S. Supreme Court again upheld the TVA Act, however, in its 1939 decision Tennessee Electric Power Company v. TVA. On August 16, 1939, TEPCO was forced to sell its assets, including Hales Bar Dam, Ocoee Dams 1 and 2, Blue Ridge Dam and Great Falls Dam to TVA for $78 million (equivalent to $ in ).
Discrimination
In 1981 the TVA Board of Directors broke with previous tradition and took a hard line against white-collar unions during contract negotiations. As a result, a class action suit was filed in 1984 in U.S. court charging the agency with sex discrimination under Title VII of the Civil Rights Act based on the large number of women in one of the pay grades negatively impacted by the new contract. An out-of-court settlement of the lawsuit was reached in 1987, in which TVA agreed to contract modifications and paid the group $5 million but admitted no wrongdoing.
Eminent domain and family removal controversies
TVA has received criticism its entire history for what some have perceived as excessive use of its authority of eminent domain and an unwillingness to compromise with landowners. All of TVA's hydroelectric projects were made possible through the use of eminent domain, and were controversial due to the more than 125,000 Tennessee Valley residents that were displaced by the agency. Residents who refused to sell their land were often forced to by court orders and lawsuits. Many of these projects also inundated historic Native American sites and early American Revolution-era settlements. Historians have criticized the TVA for forcing residents to sell their property at values less than the fair market value, and indirectly starting a unstable real estate market for farmland. In the most extreme circumstance, displaced residents would commit suicide unable to bear the events of their removal. On some occasions, land that TVA had acquired through eminent domain that was expected to be flooded by reservoirs was not flooded, and was given away to private developers.
In popular culture
The 1960 film Wild River, directed by Elia Kazan, tells the story about a family forced to relocate from their land, which has been owned by their ancestors for generations, after TVA plans to construct a dam which will flood it. While fictional, the film depicts the real-life experiences of many people forced to give up their land to TVA to make way for hydroelectric projects, and was mostly inspired by the removal of families for the Norris Project.
The 1970 James Dickey novel Deliverance, and its 1972 film adaptation focuses on four Atlanta businessmen taking a canoeing trip down a river that is being impounded by an electric utility, nodding to the TVA's early and controversial hydroelectric projects.
In the 2000 film O Brother, Where Art Thou? , the family home of the protagonist, played by George Clooney, is flooded by a reservoir constructed by the TVA. This plays a central role in the pacing of the film and the broader depression-era Mississippi context of the narrative.
"Song of the South" by country and Southern rock band Alabama features the lyrics "Papa got a job with the TVA" following the lyrics "Well momma got sick and daddy got down, The county got the farm and they moved to town" expressing the hardships and changes that southerners faced during the post recession era.
The TVA and its impact on the region are featured in the Drive-By Truckers' songs "TVA" and "Uncle Frank". In "TVA," the singer reflects on time spent with family members and a girlfriend at Wilson Dam. In "Uncle Frank", the lyrics tell the story of an unnamed hydroelectric dam being built, and the effects on the community that would become flooded upon its completion.
On November 19, 2012, Jason Isbell released a solo version of "TVA". The company still has a dominant presence in Northern Alabama, including Isbell's hometown of Muscle Shoals, as an employer and power distributor.
See also
Title 18 of the Code of Federal Regulations
Appalachian Regional Commission
Helmand and Arghandab Valley Authority, modelled on the TVA
James Bay Energy Corporation, a Crown corporation of the Quebec government for developing the James Bay Project for building various dams on rivers
List of navigation authorities in the United States
Muscle Shoals Bill
Nashville Electric Service
New Deal
New Madrid Seismic Zone
1811–12 New Madrid earthquakes
Norris, Tennessee
Tennessee Valley Authority Police
Tennessee Valley Authority v. Hill
Notes
References
Bibliography
External links
Tennessee Valley Authority in the Federal Register
WPA Photographs of TVA Archaeological Projects
Papers of Arnold R. Jones (Member of the Board of Directors, Tennessee Valley Authority), Dwight D. Eisenhower Presidential Library
TVA history
1933 establishments in Tennessee
1933 establishments in the United States
73rd United States Congress
American companies established in 1933
Appalachia
Appalachian culture
Appalachian Mountains
Companies based in Knoxville, Tennessee
Corporations chartered by the United States Congress
Economic development organizations in the United States
Electric power companies of the United States
Energy companies established in 1933
Energy companies of the United States
Energy infrastructure in Tennessee
Government agencies established in 1933
Hydroelectric power companies of the United States
Independent agencies of the United States government
New Deal agencies
New Deal in Tennessee
Nuclear power companies of the United States
Public utilities of the United States
Tennessee River
United States Department of Energy
Water management authorities in the United States |
null | null | Baton Rouge, Louisiana | eng_Latn | Baton Rouge ( ; ) is the capital of the U.S. state of Louisiana. On the eastern bank of the Mississippi River, it is the parish seat of East Baton Rouge Parish, the most-populous parish in Louisiana. Since 2020, it has been the 99th-most-populous city in the United States, and second-largest city in Louisiana after New Orleans. It is the 18th-most-populous state capital. At the U.S. Census Bureau's 2020 tabulation, it had a population of 227,470; its consolidated population was 456,781 in 2020. It is the center of the Greater Baton Rouge area, the second-largest metropolitan area in Louisiana, with a population of 870,569 as of 2020, up from 802,484 in 2010.
The Baton Rouge area owes its historical importance to its strategic site upon the Istrouma Bluff, the first natural bluff upriver from the Mississippi River Delta at the Gulf of Mexico. This allowed development of a business quarter safe from seasonal flooding. In addition, it built a levee system stretching from the bluff southward to protect the riverfront and low-lying agricultural areas. It is a culturally rich center, with settlement by immigrants from numerous European nations and African peoples brought to North America as slaves or indentured servants. It was ruled by seven different governments: French, British, and Spanish in the colonial era; the Republic of West Florida, as a United States territory and state, Confederate, and United States again since the end of the American Civil War.
Baton Rouge is a major industrial, petrochemical, medical, research, motion picture, and growing technology center of the American South. It is the location of Louisiana State University, the LSU System's flagship university and the largest institution of higher education in the state. It is also the location of Southern University, the flagship institution of the Southern University System, the only historically black college system in the United States. The Port of Greater Baton Rouge is the 10th-largest in the U.S. in terms of tonnage shipped, and is the farthest upstream Mississippi River port capable of handling Panamax ships. Major corporations participating in the economy of Baton Rouge and its metropolitan statistical area include Lamar Advertising Company, BBQGuys, Marucci Sports, Piccadilly Restaurants, Raising Cane's Chicken Fingers, ExxonMobil, and Dow Chemical Company.
History
Prehistory
Human habitation in the Baton Rouge area has been dated to 12000–6500 BC, based on evidence found along the Mississippi, Comite, and Amite rivers. Earthwork mounds were built by hunter-gatherer societies in the Middle Archaic period, from roughly the fourth millennium BC. The speakers of the Proto-Muskogean language divided into its descendant languages by about 1000 BCE; and a cultural boundary between either side of Mobile Bay and the Black Warrior River began to appear between about 1200 BCE and 500 BCE, a period called the Middle "Gulf Formational Stage". The Eastern Muskogean language began to diversify internally in the first half of the first millennium AD.
The early Muskogean societies were the bearers of the Mississippian culture, which formed around 800 AD and extended in a vast network across the Mississippi and Ohio valleys, with numerous chiefdoms in the Southeast, as well. By the time the Spanish made their first forays inland from the shores of the Gulf of Mexico in the early 16th century, by some evidence many political centers of the Mississippians were already in decline, or abandoned. At the time, this region appeared to have been occupied by a collection of moderately sized native chiefdoms, interspersed with autonomous villages and tribal groups. Other evidence indicates these Mississippian settlements were thriving at the time of the first Spanish contact. Later Spanish expeditions encountered the remains of groups who had lost many people and been disrupted in the aftermath of infectious diseases, chronic among Europeans, unknowingly introduced by the first expedition.
Colonial period
French explorer Pierre Le Moyne d'Iberville led an exploration party up the Mississippi River in 1698. The explorers saw a red pole marking the boundary between the Houma and Bayagoula tribal hunting grounds. The French name le bâton rouge ("the red stick") is the translation of a native term rendered as Istrouma, possibly a corruption of the Choctaw iti humma ("red pole"); André-Joseph Pénicaut, a carpenter traveling with d'Iberville, published the first full-length account of the expedition in 1723. According to Pénicaut,
From there [Manchacq] we went five leagues higher and found very high banks called écorts in that region, and in savage called Istrouma which means red stick [bâton rouge], as at this place there is a post painted red that the savages have sunk there to mark the land line between the two nations, namely: the land of the Bayagoulas which they were leaving and the land of another nation—thirty leagues upstream from the baton rouge—named the Oumas.The location of the red pole was presumably at Scott's Bluff, on what is now the campus of Southern University. It was reportedly a painted pole adorned with fish bones.
The settlement of Baton Rouge by Europeans began in 1721 when French colonists established a military and trading post. Since European settlement, Baton Rouge has been governed by France, Britain, Spain, Louisiana, the Republic of West Florida, the United States, the Confederate States, and the United States again. In 1755, when French-speaking settlers of Acadia in Canada's Maritime provinces were expelled by British forces, many took up residence in rural Louisiana. Popularly known as Cajuns, the descendants of the Acadians maintained a separate culture. During the first half of the 19th century, Baton Rouge grew steadily as the result of steamboat trade and transportation.
Modern history
Baton Rouge was incorporated in 1817. In 1822, the Pentagon Barracks complex of buildings was completed. The site has been used by the Spanish, French, British, Confederate States Army, and United States Army and was part of the short-lived Republic of West Florida. In 1951, ownership of the barracks was transferred to the state of Louisiana. In 1976, the complex was listed on the National Register of Historic Places.
Acquisition of Louisiana by the United States in 1803 was a catalyst for increased Anglo-American settlement, especially in the northern part of the state. In 1846, the state legislature designated Baton Rouge as Louisiana's new capital to replace "sinful" New Orleans. The architect James Dakin was hired to design the old Louisiana State Capitol, with construction beginning in late 1847. Rather than mimic the United States Capitol, as many other states had done, he designed a capitol in Neo-Gothic style, complete with turrets and crenellations, and stained glass; it overlooks the Mississippi. It has been described as the "most distinguished example of Gothic Revival" architecture in the state and has been designated as a National Historic Landmark.
By the outbreak of the American Civil War, the population of Baton Rouge was nearly 5,500. The war nearly halted economic progress, except for businesses associated with supplying the Union Army occupation of the city, which began in the spring of 1862 and lasted for the duration of the war. The Confederates at first consolidated their forces elsewhere, during which time the state government was moved to Opelousas and later Shreveport. In the summer of 1862, about 2,600 Confederate troops under generals John C. Breckinridge (the former Vice President of the United States) and Daniel Ruggles attempted to recapture Baton Rouge.
After the war, New Orleans temporarily served as the seat of the Reconstruction era state government. When the Bourbon Democrats regained power in 1882, after considerable intimidation and voter suppression of black Republicans, they returned the state government to Baton Rouge, where it has since remained. In his 1893 guidebook, Karl Baedeker described Baton Rouge as "the Capital of Louisiana, a quaint old place with 10,378 inhabitants, on a bluff above the Mississippi".
In the 1950s and 1960s, the petrochemical industry had a boom in Baton Rouge, stimulating the city's expansion beyond its original center. The changing market in the oil business has produced fluctuations in the industry, affecting employment in the city and area.
A building boom began in the city in the 1990s and continued into the 2000s, during which Baton Rouge was one of the fastest-growing cities in the South in terms of technology. Metropolitan Baton Rouge was ranked as one of the fastest-growing metropolitan areas in the U.S. (with a population under 1 million), with 602,894 in 2000 and 802,484 people as of the 2010 U.S. census. After the extensive damage in New Orleans and along the coast from Hurricane Katrina on August 29, 2005, this city accepted as many as 200,000 displaced residents.
In 2010, Baton Rouge started a market push to become a test city for Google's new super high speed fiber optic line known as GeauxFiBR.
In 2016, the Greater Baton Rouge metropolitan area was heavily affected by the shooting of police officers in July and flooding in August.
Geography
Baton Rouge is located on the banks of the Mississippi River in southeastern Louisiana's Florida Parishes region. The city of Baton Rouge is approximately from New Orleans, from Alexandria, from Lafayette and from Shreveport. It is also from Jackson, Mississippi and from Houston, Texas. Baton Rouge lies on a low elevation of 56 to a little over 62 feet above sea level. The city's proximity to Greater New Orleans, Greater Jackson, and Greater Houston makes it a prominent transportation hub between the metropolitan regions.
The city of Baton Rouge is the capital of Louisiana and the parish seat of East Baton Rouge Parish. According to the United States Census Bureau, the city has a total area , of which are land and (2.81%) are covered by water. The city is located on the first set of bluffs north of the Mississippi River Delta's coastal plains. Because of its prominent location along the river and on the bluffs, which prevents flooding, the French built a fort in the city in 1719. Baton Rouge is the third-southernmost capital city in the continental United States, after Austin, Texas, and Tallahassee, Florida. It is the cultural and economic center of the Greater Baton Rouge metropolitan area.
Climate
Baton Rouge has a humid subtropical climate (Köppen Cfa), with mild winters, hot and humid summers, moderate to heavy rainfall, and the possibility of damaging winds and tornadoes yearlong. The area's average precipitation is 55.55 inches (141.1 cm) of rain and 0.1 inches (0.25 cm) of snow annually. With ample precipitation, Baton Rouge is fifth on the list of wettest cities in the United States. Snow in the Baton Rouge area is usually rare, although it snowed in three consecutive years recently: December 11, 2008, December 4, 2009, and February 12, 2010. The yearly average temperature for Baton Rouge is while the average temperature for January is and July is . The area is usually free from extremes in temperature, with some cold winter fronts, but those are usually brief.
Baton Rouge's proximity to the Gulf of Mexico exposes the metropolitan region to hurricanes. On September 1, 2008, Hurricane Gustav struck the city and became the worst hurricane ever to hit the Baton Rouge area. Winds topped , knocking down trees and powerlines and making roads impassable. The roofs of many buildings suffered tree damage, especially in the Highland Road, Garden District, and Goodwood areas. The city was shut down for five days and a curfew was put in effect. Rooftop shingles were ripped off, signs blew down, and minor structural damage occurred.
Demographics
The U.S. Census Bureau determined Baton Rouge had a total population of 227,470 in 2020, down from 229,493 in 2010. The 2019 American Community Survey estimated 443,763 people lived in the consolidated area of Baton Rouge (East Baton Rouge Parish). The metropolitan population increased to 3.6% as a result of suburbanization in 2019, to an estimated 854,884. In 2020, the metropolitan population was 870,569. In 2019, there were 83,733 households with an average of 2.58 people per household. Baton Rouge had a population density of 2,982.5 people per square mile.
Per the 2019 census estimates, 21.6% of households had children under the age of 18 living in them. The owner-occupied housing rate of Baton Rouge was 49.8% and the median value of an owner-occupied housing unit was $174,000. The median monthly owner-costs with a mortgage were $1,330 and the cost without a mortgage was $382. Baton Rouge had a median gross rent of $879, making it one of the Southern U.S.'s most affordable major cities. In the city, the median household income was $44,470 and the per capita income was $28,491 at the 2019 census estimates. Roughly 24.8% of the city lived at or below the poverty line.
At the 2010 United States census, there were 88,973 households and 52,672 families. Of all households in 2010, 28.1% had children under the age of 18, 35.8% were married couples living together, 19.0% had a female householder with no husband present, and 40.8% were not families. About 31.7% of all households were made up of individuals, and 8.6% had someone living alone who was 65 years of age or older. The average household size was 2.42 and the average family size was 3.12. In the city, the population was distributed as 24.4% under the age of 18, 17.5% from 18 to 24, 27.2% from 25 to 44, 19.4% from 45 to 64, and 11.4% who were 65 years of age or older.
The median age in Baton Rouge in 2010 was 30 years. For every 100 females, there were 90.5 males. For every 100 females age 18 and over, there were 86.3 males. The median income for a household in the city was $30,368, and for a family was $40,266. Males had a median income of $34,893 versus $23,115 for females. The per capita income for the city was $18,512. About 18.0% of families and 24.0% of the population were below the poverty line, including 31.4% of those under age 18 and 13.8% of those ages 65 or over.
Race and ethnicity
At the 2010 United States census, the racial and ethnic makeup of Baton Rouge proper was 54.54% Black and African American, 39.37% White American, 0.5% American Indian and Alaska Native, 3.5% Asian, and 1.3% from two or more races. Hispanics or Latin Americans were 3.5% of the population. Non-Hispanic Whites were 37.8% of the population, down from 70.5% in 1970.
In 2019, the racial and ethnic makeup was 36.6% non-Hispanic white, 54.7% Black and African American, 0.3% American Indian and Alaska Native, 3.5% Asian American, 1.3% from two or more races, and 3.7% Hispanic or Latin American of any race. Of the population, roughly 5.5% were foreign-born from 2015 to 2019. The 2014 census estimates determined the Black and African American population was 56.1%, non-Hispanic whites 35.7%, American Indians and Alaska Natives 0.2%, Asians 2.8%, multiracial 1.1%, and Hispanics or Latin Americans 4.0%. Since the 2010 U.S. census, the non-Hispanic white and Black populations have fluctuated, while Black and African American residents remain the largest group in the area, similar to the trends in New Orleans.
Religion
Christianity is the predominant religion practiced in the Baton Rouge area. There is a large Catholic influence in the city and metro area (22.6%), owing in part to French and Spanish settlement, while Baptists are the second largest denomination (20.0%). The Catholic population are primarily served by the Latin Church's Roman Catholic Diocese of Baton Rouge. Prominent Baptist denominations include the National Baptist Convention (USA), National Baptist Convention of America, the Progressive National Baptist Convention, the American Baptist Churches USA, and Southern Baptist Convention.
Other large Christian bodies in the area include Methodists, Anglicans or Episcopalians, Pentecostals, Presbyterians, Latter-Day Saints, and Lutherans. Christians including Jehovah's Witnesses, the Metropolitan Community Church, Christian Unitarians, and the Eastern Orthodox among others collectively make up 14% of the study's other Christian demographic. Notable Methodist and Anglican/Episcopalian jurisdictions operating throughout the Greater Baton Rouge area include the United Methodist Church, African Methodist Episcopal Church, the Episcopal Diocese of Louisiana, and the Anglican Church in North America. Baton Rouge's Pentecostals mainly affiliate with the Assemblies of God USA and the Church of God in Christ, and Presbyterians are mainly members of the Presbyterian Church (USA).
The second-largest religion in Baton Rouge and its metropolitan area is Islam (0.4%). There are currently over six mosques in the Baton Rouge area, primarily affiliated with Sunni Islam. The Nation of Islam is also another prominent branch of the religion practiced. The Muslim demographic has grown out of Middle Eastern immigration and African American Muslim missionary work. The first Islamic private school in Baton Rouge was established in 2019. As of 2019, Orthodox Jews make up 0.2% of Baton Rouge's religious population, and 0.6% identify with eastern faiths including Buddhism and Hinduism. New religious movements including contemporary paganism have small communities in the area, and a minority affiliate with Haitian Vodou, Louisiana Voodoo, and Hoodoo. The remainder of Baton Rouge's population is spiritual but not religious, agnostic, or atheist.
Economy
Baton Rouge enjoys a strong economy that has helped the city be ranked as one of the "Top 10 Places for Young Adults" in 2010 by portfolio.com and one of the top 20 cities in North America for economic strength by the Brookings Institution. In 2009, the city was ranked by CNN as the 9th-best place in the country to start a new business. Lamar Advertising Company has its headquarters in Baton Rouge. Other notable companies headquartered in the city include BBQGuys, Marucci Sports, Piccadilly Restaurants, and Raising Cane's Chicken Fingers.
Baton Rouge is the furthest inland port on the Mississippi River that can accommodate ocean-going tankers and cargo carriers. The ships transfer their cargo (grain, oil, cars, containers) at Baton Rouge onto rails and pipelines (to travel east–west) or barges (to travel north). Deep-draft vessels cannot pass the Old Huey Long Bridge because the clearance is insufficient. In addition, the river depth decreases significantly just to the north, near Port Hudson.
Baton Rouge's largest industry is petrochemical production and manufacturing. ExxonMobil's Baton Rouge Refinery complex is the fourth-largest oil refinery in the country; it is the world's 10th largest. Baton Rouge also has rail, highway, pipeline, and deep-water access. Dow Chemical Company has a large plant in Iberville Parish near Plaquemine, south of Baton Rouge. Shaw Construction, Turner, and Harmony all started with performing construction work at these plants.
In addition to being the state capital and parish seat, the city is the home of Louisiana State University, which employs over 5,000 academic staff. One of the largest single employers in Baton Rouge is the state government, which consolidated all branches of state government downtown at the Capitol Park complex.
The city has a substantial medical research and clinical presence. Research hospitals have included Our Lady of the Lake, Our Lady of the Lake Children's Hospital (affiliated with St. Jude Children's Research Hospital), Mary Bird Perkins Cancer Center, and Earl K. Long (closed 2013). Together with an emerging medical corridor at Essen Lane, Summa Avenue and Bluebonnet Boulevard, Baton Rouge is developing a medical district expected to be similar to the Texas Medical Center. LSU and Tulane University have both announced plans to construct satellite medical campuses in Baton Rouge to partner with Our Lady of the Lake Medical Center and Baton Rouge General Medical Center, respectively.
Southeastern Louisiana University and Franciscan Missionaries of Our Lady University both have nursing schools in the medical district off Essen Lane. Louisiana State University's Pennington Biomedical Research Center, which conducts clinical and biological research, also contributes to research-related employment in the area around the Baton Rouge medical district.
The film industry in Louisiana has increased dramatically since the beginning of the 21st century, aided by generous tax incentives adopted by the state in 2002. In September 2013, the Baton Rouge Film Commission reported that the industry had brought more than $90 million into the local economy in 2013. Baton Rouge's largest production facility is the Celtic Media Centre, opened in 2006 by a local group in collaboration with Raleigh Studios of Los Angeles. Raleigh dropped its involvement in 2014.
Culture and arts
Baton Rouge is the middle ground of South Louisiana cultures, having a mix of Cajun and Creole Catholics and Baptists of the Florida Parishes and South Mississippi. Baton Rouge is a “college town” with Baton Rouge Community College, Louisiana State University, Franciscan Missionaries of Our Lady University, and Southern University, whose students make up some 20% of the city population. In a sizable international population of over 11,300, the largest groups were people of Hispanic and Latino or Vietnamese descent. This contributes to Baton Rouge's unique culture and diversity.
Arts and theater
Baton Rouge has an expanding visual arts scene, which is centered downtown. Professional performing arts organizations include Theatre Baton Rouge, the Baton Rouge Symphony Orchestra, Baton Rouge Ballet Theatre, Opera Louisiane and Playmakers, a professional Children’s Theatre. This increasing collection of venues includes the Shaw Center for the Arts. Opened in 2005, the facility houses the Brunner Gallery, LSU Museum of Art, the Manship Theatre, a contemporary art gallery, traveling exhibits, and several eateries. Another prominent facility is the Louisiana Art and Science Museum, which contains the Irene W. Pennington Planetarium, traveling art exhibits, space displays, and an ancient Egyptian section. Several smaller art galleries, including the Baton Rouge Gallery, offering a range of local art, are scattered throughout the city.
The city has several designated arts and cultural districts, the most prominent of which are the Mid-City Cultural District and the Perkins Road Arts District. These districts provide tax incentives, mostly in the form of exempting state tax on purchases, to promote cultural activity in these areas.
A performing arts scene is emerging. LSU's Swine Palace is the foremost theatre company in the city, largely made up of students of LSU's MFA acting program, as well as professional actors and stage managers. A group of physical theatre and circus artists from LSU traveled to Edinburgh, Scotland, in summer 2012 to perform Dante in what has become the world's largest Fringe Festival. The show ran in Baton Rouge before going to Fringe, and featured movement, acrobatics, and aerial silk.
Theatre Baton Rouge offers a diverse selection of live theatre performances. Opera Louisiane is Baton Rouge's only professional opera company. The Baton Rouge Ballet Theatre is Baton Rouge's professional ballet company. The Nutcracker– A Tale from the Bayou sets the familiar holiday classic in 19th-century Louisiana and has become a Baton Rouge holiday tradition. A Tale from the Bayou features professional dancers, a live orchestra, and more than 300 area children.
Baton Rouge is also home to Forward Arts, a youth writing organization. Forward Arts won the international youth poetry slam, Brave New Voices in 2017, and was the first team from the Southern United States to ever win the competition. Forward Arts is the only youth spoken-word organization in the state of Louisiana. It was founded by Dr. Anna West in 2005, and first housed in the Big Buddy Program.
Baton Rouge is also home to Of Moving Colors Productions, the premier contemporary dance company in the city. For more than 30 years it has brought in internationally established choreographers to create stunning performances. In addition, they conduct extensive community outreach to children and young adults.
Performing venues include the Baton Rouge River Center, Baton Rouge River Center Theatre for the Performing Arts, which seats about 1,900; the Manship Theatre, which is located in the Shaw Center for the Arts and seats 350; and the Reilly Theater, which is home to Swine Palace, a nonprofit professional theater company associated with the Louisiana State University Department of Theatre.
The Baton Rouge Symphony Orchestra has operated since 1947 and currently performs at the River Center Music Hall downtown. Today, it presents more than 60 concerts annually, directed by Timothy Muffitt and David Torns. The BRSO's educational component, the Louisiana Youth Orchestra, made its debut in 1984. It includes almost 180 musicians under the age of 20.
Miss USA pageants
Baton Rouge was chosen to host the Miss USA 2014 Pageant. It took over downtown Baton Rouge as Nia Sanchez, Miss Nevada USA, took home the crown, with Miss Louisiana USA Brittany Guidry coming close to the win with the third runner-up spot and fourth overall. Veteran pageant host Giuliana Rancic and MSNBC news anchor Thomas Roberts introduced the 51 contestants; there were 20 semifinalists. Cosmo weighed in on the contest, complimenting home state girl Miss Louisiana. Celebrity judges included actress Rumer Willis, NBA star Karl Malone, singer Lance Bass, and actor Ian Ziering. In the contest's 62-year history, this was the first year that viewers got to vote to keep one of their favorite contestants in the top six by tweeting the hashtag #SaveTheQueen. Baton Rouge hosted Miss USA 2015 again on July 12, 2015, which was won by actress and Miss Oklahoma USA Olivia Jordan. Baton Rouge was also the site of the 2005 Miss Teen USA Pageant.
Tourism and recreation
The many architectural points of interest in Baton Rouge range from antebellum to modern. The neo-gothic Old Louisiana State Capitol was built in the 1850s as the first state house in Baton Rouge. It was later replaced by the 450-ft-tall, art deco New Louisiana State Capitol, which was the tallest building in the South when it was completed. Several plantation homes in the area, such as Magnolia Mound Plantation House, Myrtles Plantation, and Nottoway Plantation, showcase antebellum-era architecture.
Louisiana State University has more than 250 buildings designed in Italian Renaissance style, one of the nation's largest college stadiums, and is endowed with many live oaks. The downtown has several examples of modern and contemporary buildings, including the Capitol Park Museum.
A number of structures, including the Baton Rouge River Center, Louisiana State Library, LSU Student Union, Louisiana Naval Museum, Bluebonnet Swamp Interpretive Center, Louisiana Arts and Sciences Center, Louisiana State Archive and Research Library, and the Pennington Biomedical Research Center, were designed by local architect John Desmond. The Pentagon Barracks Museum and Visitors Center is located within the barracks complex and the Yazoo and Mississippi Valley Railroad Company Depot, currently houses the Louisiana Art and Science Museum.
Museums around town offer a variety of genres. The Capitol Park Museum and the Old Louisiana State Capitol Museum display information on state history and have many interactive exhibits. The Shaw Center for the Arts and the Louisiana Art and Science Museum showcase varied arts. LASM also includes science exhibits and a planetarium. Other museums include the LSU Museum of Natural Science and the USS Kidd. The Odell S. Williams Now And Then African-American Museum chronicles the progression and growth of African-American.
Other attractions include the Mall of Louisiana and Perkins Rowe, amusement parks of Dixie Landin'/Blue Bayou, and dining at the Louisiana-cuisine restaurants.
The 2nd Annual Slam’d & Cam’d Car Show, slated for July 10, 2021, was scheduled to begin after the cancellation of the 2020 show due to the pandemic. The 2021 show planned to present than 300 muscle cars and include TV personality Courtney Hansen and MotorTrend Group's, Iron Resurrection stars, Joe and Amanda Martin.
Sports
College sports play a major role in the culture of Baton Rouge. The LSU Tigers and the Southern University Jaguars are NCAA Division I athletic programs with the LSU Tigers football and Southern Jaguars football teams being the local college American football teams. College baseball, basketball, and gymnastics are also popular.
Much of the city's sport's attention is focused on the professional teams in Greater New Orleans. Baton Rouge has had multiple minor-league baseball teams (the Baton Rouge Red Sticks), soccer teams (Baton Rouge Bombers), indoor football teams, a basketball team, and a hockey team (Baton Rouge Kingfish). The Baton Rouge Rugby Football Club or Baton Rouge Redfish 7, which began playing in 1977, has won numerous conference championships. Currently, the team competes in the Deep South Rugby Football Union. It also has an Australian rules football team, the Baton Rouge Tigers, which began playing in 2004 and competes in the USAFL. In addition, Baton Rouge is home to Red Stick Roller Derby, a WFTDA Division 3 roller derby league. Baton Rouge is also home to the Baton Rouge Soccer Club in the Gulf Coast Premier League and the Baton Rouge Rougarou, a college summer league baseball team and member of the Texas Collegiate League.
Parks and recreation
Baton Rouge has an extensive park collection run through the Recreation & Park Commission for the Parish of East Baton Rouge (BREC). The largest park is City Park near the Louisiana State University flagship campus. The Baton Rouge Zoo is also run through BREC and includes over 1800 species.
National protected areas
Atchafalaya National Heritage Area
Baton Rouge National Cemetery
Independence Park Botanic Gardens
Laurens Henry Cohn, Sr. Memorial Plant Arboretum
LSU Hilltop Arboretum
Magnolia Cemetery
Port Hudson National Cemetery
Government
The city of Baton Rouge and the Parish of East Baton Rouge have been run by a consolidated government since 1947. It combined the City of Baton Rouge government with the rural areas of the parish, allowing people outside the limits of the City of Baton Rouge to use city services. Though the city and parish have a consolidated government, this differs slightly from a traditional consolidated city-county government. The cities of Zachary, Baker, and Central continue to operate their own individual city governments within East Baton Rouge Parish. Under this system, Baton Rouge has the uncommon office of "mayor-president", which consolidates the executive offices of "mayor of Baton Rouge" and "president of East Baton Rouge Parish". Though Zachary, Baker, and Central each have their own individual mayors, citizens living in these three municipalities are still a part of the constituency who can vote and run in elections for mayor-president and metro council.
The mayor-president's duties include setting the agenda for the government and managing the government's day-to-day functions. They are also responsible for supervising departments, as well as appointing the department heads. The mayor does not set the city's public policy because that is the role of the Metropolitan Council. However, the mayor-president does have some influence on the policy through appointments and relationships with Council members.
The current Mayor-President of Baton Rouge is Sharon Weston Broome, a former Louisiana State Legislator. Broome, a Democrat, succeeded Kip Holden, also a Democrat, as Mayor-President on January 2, 2017, after defeating Bodi White in a close runoff on December 10, 2016. She served in the Louisiana House of Representatives from 1992 to 2004, and in the Louisiana State Senate from 2004 to 2016. She was elected by the senate to serve as the Senate President Pro Tempore from 2008 to 2016.
Metropolitan Council
When the city and parish combined government, the city and parish councils consolidated to form the East Baton Rouge Parish Metropolitan Council. The Metro Council is the legislative branch of the Baton Rouge government. Its 12 district council members are elected from single-member districts. They elect from among themselves the mayor-president pro tempore. The Mayor-President Pro Tempore presides over the council's meetings and assumes the role of the Mayor-President if the Mayor-President is unable to serve. The council members serve four-year terms and can hold office for three terms.
In the late 1960s, Joe Delpit, a local African-American businessman who owns the successful and still operating Chicken Shack, was elected as the first black council member in Baton Rouge. As in other cities of Louisiana and the South, African Americans had been largely disenfranchised for decades into the 20th century. The Chicken Shack, with multiple locations, in 2015 was reported as the oldest continually operating business in Baton Rouge.
The Metro Council's main responsibilities are setting the policy for the government, voting on legislation, and approving the city's budget. The Council makes policies for the following: the City and Parish General Funds, all districts created by the council, the Greater Baton Rouge Airport District, the Public Transportation Commission, the East Baton Rouge Parish Sewerage Control Commission and the Greater Baton Rouge Parking Authority.
Education
Baton Rouge is stimulated by many universities. Louisiana State University and Agricultural and Mechanical College, generally known as Louisiana State University or LSU, is a public, coeducational university that is the flagship campus of the Louisiana State University System. LSU is the largest university in Louisiana with over 30,000 students and 1,300 full-time faculty members. Southern University and A&M College, generally known as Southern University or SU, is the flagship institution of the Southern University System, the only historically black land-grant university system in the United States. SU is the largest HBCU and second-oldest public university in Louisiana.
Virginia College opened in October 2010 and offers students training in areas such as cosmetology, business, health, and medical billing. Franciscan Missionaries of Our Lady University is an independent Catholic institution also in the Baton Rouge medical district that has programs in nursing, health sciences, humanities, behavioral sciences, and arts and sciences. It has an associated hospital, Our Lady of the Lake Regional Medical Center. Tulane University is also opening a satellite medical school at Baton Rouge General's Mid City Campus in 2011.
Southeastern Louisiana University School of Nursing is located in the medical district on Essen Lane in Baton Rouge. Southeastern offers traditional baccalaureate and master's degree programs, as well as LPN and RN to BSN articulation. Baton Rouge Community College is an open-admission, two-year postsecondary public community college, established on June 28, 1995. The college settled into a permanent location in 1998. The college's current enrollment is more than 8,000 students. The Pennington Biomedical Research Center houses 48 laboratories and 19 core research facilities.
Primary and secondary schools
East Baton Rouge Parish Public Schools operates primary and secondary schools serving the city. The city of Baton Rouge is also home to 27 charter schools with a total enrollment of an estimated 11,000 students as of 2020. One of the latest includes the Mentorship Academy in downtown Baton Rouge, which leverages its location downtown to establish internship opportunities with local businesses as well as provide a high-tech classroom environment to focus on a digital animation curriculum.
The East Baton Rouge Parish School System is the second-largest public school system in the state and contains nine U.S. Blue Ribbon schools and a nationally renowned Magnet program. The school system serves more than 42,850 students and with the help of 6,250 teachers and faculty, the district has shown growth and increase in its District Performance Score. The East Baton Rouge Parish Public Schools serve East Baton Rouge Parish and has 90 schools with 56 elementary schools, 16 middle schools, and 18 high schools.
Libraries
The State Library of Louisiana is in Baton Rouge. The Louisiana Legislature created the Louisiana Library Commission in 1920. This later became the State Library of Louisiana. The State Library provides Louisiana residents with millions of items with its collections, electronic resources, and the statewide network for lending.
The East Baton Rouge Parish Library System has 14 local libraries with one main library and 13 community libraries. The main library at Goodwood houses genealogy and local history archives. The library system is an entity of the city-parish government. The system has been in operation since 1939. It is governed by the EBR Parish government and directed by the Library Board of Control. The Baton Rouge Metropolitan Council appoints the seven-member board and then the board appoints a director. According to its website, all branches are open seven days a week to assist the public with reference and information and computer access.
The Louisiana State Archives' Main Research Library is located in Baton Rouge, as well. It houses general history books, census indices, immigration schedules, church records, and family histories. The library also has a computerized database of more than two million names that has various information about these people including census, marriage, and social security filing information.
Louisiana State University and the Louisiana State University Law Center have libraries on their respective Baton Rouge campuses. Southern University and A&M College and the Southern University Law Center also have libraries on their respective Baton Rouge campuses.
Media
The major daily newspaper for the Greater Baton Rouge metropolitan area is The Advocate, publishing since 1925. Prior to October 1991, Baton Rouge also had an evening newspaper, The State-Times—at that time, the morning paper was known as The Morning Advocate. Other publications include: Baton Rouge Parents Magazine, Pink & Blue Magazine, The Daily Reveille, The Southern Review, 225 magazine, DIG, Greater Baton Rouge Business Report, inRegister magazine, 10/12 magazine, Country Roads magazine, 225Alive, Healthcare Journal of Baton Rouge, Southern University Digest, and The South Baton Rouge Journal.
Other newspapers in East Baton Rouge Parish include the Central City News and The Zachary Post.
The Greater Baton Rouge area is well served by television and radio. The market is the 95th-largest designated market area in the U.S. Major television network affiliates serving the area include:
2 WBRZ-TV (ABC)
9 WAFB (CBS)
20 KZUP-CD (RTN)
21 WBRL-CD (The CW)
27 WLPB (PBS/LPB)
30 WLFT-CD (MeTV)
33 WVLA (NBC)
36 KBTR-CD (This TV)
39 WBXH-CD (My Network TV)
44 WGMB (Fox)
Baton Rouge also offer local government-access television-only channels on Cox Cable channel 21.
Infrastructure
Health and medicine
Baton Rouge is served by several hospitals and clinics:
Baton Rouge General Medical Center – Mid-City Campus – 3600 Florida Boulevard
Baton Rouge General Medical Center – Bluebonnet Campus – 8585 Picardy Avenue
HealthSouth Rehabilitation Hospital – 8595 United Plaza Boulevard
Our Lady of the Lake Regional Medical Center – 5000 Hennessy Boulevard
Ochsner Medical Center – 1700 Medical Center Drive
Our Lady of the Lake Children's Hospital – 5000 Hennessy Boulevard
Ochsner Medical Complex – The Grove – 10310 The Grove Boulevard
Communication
Most of the Baton Rouge area's high-speed internet, broadband, and fiber optic communications are provided by Eatel, AT&T Inc., Charter Communications, or Cox Communications. In 2006, Cox Communications linked its Lafayette, Baton Rouge, and New Orleans markets with fiber-optic infrastructure. Other providers soon followed suit, and fiber optics have thus far proven reliable in all hurricanes since they were installed, even when mobile and broadband services are disrupted during storms. In 2001, the Supermike computer at Louisiana State University was ranked as the number-one computer cluster in the world, and remains one of the top 500 computing sites in the world.
Military installations
Baton Rouge is home station to the Louisiana Army National Guard 769th Engineer Battalion, which recently had units deployed to Iraq and Afghanistan. The armory located near LSU has three company-sized units: 769th HSC (headquarters support company); 769th FSC (forward support company); and the 927th Sapper Company. Other units of the battalion are located at Napoleonville (928th Sapper Company); Baker, Louisiana (926th MAC mobility augmentation company); and Gonzales, Louisiana (922nd Horizontal Construction Company).
The 769th Engineer Battalion is part of the 225th Engineer Brigade, which is headquartered in Pineville, Louisiana, at Camp Beauregard. Four engineer battalions and an independent bridging company are in the 225th Engineer Brigade, which makes it the largest engineer group in the US Army Corps of Engineers.
Baton Rouge is also home to 3rd Battalion, 23rd Marine Regiment (3/23), a reserve infantry battalion in the United States Marine Corps located throughout the Midwestern United States consisting of about 800 marines and sailors. The battalion was first formed in 1943 for service in the Pacific Theater of Operations during World War II, taking part in a number of significant battles including those at Saipan and Iwo Jima before being deactivated at the end of the war. In the early 1960s, the unit was reactivated as a reserve battalion. The battalion is headquartered in Saint Louis, Missouri, with outlying units throughout the Midwestern United States. 3/23 falls under the command of the 23rd Marine Regiment and the 4th Marine Division. Recent operations have included tours in Iraq and Afghanistan.
Transportation
Shipping
The Port of Baton Rouge is the 9th-largest in the United States in terms of tonnage shipped, and is the farthest upstream Mississippi River port capable of handling Panamax ships.
Highways and roads
Interstates
Baton Rouge has three interstate highways: I-10, I-12
(Republic of West Florida Parkway), and I-110 (Martin Luther King Jr. Expressway).
Interstate 10 enters the city from the Horace Wilkinson Bridge over the Mississippi River, curving at an interchange with Interstate 110 southeast, crossing the LSU lakes and Garden District before reaching an interchange with I-12 (referred to as the 10/12 split). It curves further southeast, towards New Orleans as it crosses Essen Lane near the Medical District. It passes Bluebonnet Blvd and the Mall of Louisiana at exit 162, and leaves Baton Rouge after interchanges with Siegen Lane and Highland Road.
Interstate 12 (The Republic of West Florida Parkway) begins in the city at the I-10/I-12 split east of College Drive, and goes east from there, crossing Essen Lane, Airline Hwy, Sherwood Forest Blvd, Millerville Road, and O'neal Lane before leaving the city when crossing the Amite River.
Interstate 110 (The Martin Luther King Jr. Expressway) stretches 8 miles in a north–south direction from the east end of the Horace Wilkinson Bridge to Scenic Highway in Scotlandville, Louisiana. It passes through downtown, North Baton Rouge, and Baton Rouge Metro Airport before ending at Scenic Highway.
US highways and major roads
Baton Rouge has two US highways, along with their business counterparts: Airline Highway (US 61) and Florida Boulevard.
US 190 enters the city from the Huey P. Long Bridge, beginning a concurrency with US 61 after an interchange with Scenic Highway, near Scotlandville. Its name is Airline Highway from this interchange to the interchange with Florida Blvd. At this interchange, US 190 turns east to follow Florida Blvd through Northeast Baton Rouge, exiting the city at the Amite River.
US 61 enters Baton Rouge as Scenic Highway until it reaches Airline Highway (US 190). It becomes concurrent with US 190 until Florida Blvd, where it continues south, still called Airline Highway. It passes through Goodwood and Broadmoor before an interchange with I-12. It continues southeast past Bluebonnet Blvd/Coursey Blvd, Jefferson Hwy, and Sherwood Forest Blvd/Siegen Lane before exiting the city at Bayou Manchac.
US 61/190 Business runs west along Florida Boulevard (known as Florida Street from Downtown east to Mid City) from Airline Highway to River Road in downtown. The cosigned routes run from Florida St. north along River Road, passing the Louisiana State Capitol and Capitol Park Complex before intersecting with Choctaw Drive. North of this intersection River Road becomes Chippewa Street and curves to the East. US 61/190 Business leaves Chippewa Street at its intersection with Scenic Highway. The route follows Scenic Highway to Airline Highway, where it ends. North of Airline on Scenic and East of Scenic Highway on Airline is US 61. US 190 is East and West of Scenic on Airline Highway.
These are important surface streets with designated state highway numbers: Greenwell Springs Road (LA 37), Plank Road/22nd Street (LA 67), Burbank Drive/Highland Road (LA 42), Nicholson Drive (LA 30), Jefferson Highway/Government Street (LA 73), Scotlandville/Baker/Zachary Highway (LA 19), Essen Lane (LA 3064), Bluebonnet Blvd/Coursey Blvd (LA 1248), Siegen Lane (LA 3246), and Perkins Road/Acadian Thruway (LA 427).
Traffic issues and highway upgrades
According to the 2008 INRIX National Traffic Scorecard, which ranks the top 100 congested metropolitan areas in the U.S., Baton Rouge is the 33rd-most congested metro area in the country. However, at a population rank of 67 out of 100, it has the second-highest ratio of population rank to congestion rank, higher than even the Los Angeles-Long Beach-Santa Ana metropolitan area, indicating a remarkably high level of congestion for the comparatively low population. According to the Scorecard, Baton Rouge was the only area out of all 100 to show an increase in congestion from 2007 to 2008 (+ 6%). The city also tied for the highest jump in congestion rank over the same period (14 places).
Interstate 12 used to have a major bottleneck at O'Neal Lane. The interstate was three lanes wide in each direction to the O'Neal Lane exit, where the interstate abruptly became two lanes in each direction and crossed the narrow Amite River Bridge. This stretch of road, called "a deathtrap" by one lawmaker, had become notorious for traffic accidents, many with fatalities. In 2007, ten people died in traffic accidents within a three-month period on this section of road. Governor Bobby Jindal and the Baton Rouge legislative delegation, in 2009, were successful in allocating state and federal funding to widen I-12 from O'neal Lane to Range Avenue (Exit 10) in Denham Springs. The construction was completed in mid-2012 and has significantly improved the flow of traffic. In 2010, The American Reinvestment and Recovery Act provided committed federal funds to widen I-12 from the Range Avenue Exit to Walker, Louisiana. Noticing the significant improvement in commute times, Bobby Jindal further funded widening to Satsuma, Louisiana.
Interstate 10 West at Bluebonnet Road also ranked within the top 1000 bottlenecks for 2008, and I-10 East at Essen Lane and Nicholson Drive ranked not far out of the top 1000. A new exit to the Mall of Louisiana was created in 2006, and the interstate was widened between Bluebonnet Blvd and Siegen Lane. However, the stretch of I-10 from the I-10/I-12 split to Bluebonnet Blvd was not part of these improvements and remained heavily congested during peak hours. In response, a widening project totaling at least $87 million began in late 2008. Interstate 10 was widened to three lanes over a five-year period between the I-10/I-12 split and Highland Road. In 2010, the American Reinvestment and Recovery Act provided supplemental funding for this project to extend to the Highland Road exit in East Baton Rouge Parish. Commute times have since plummeted for this section of interstate.
Surface streets in Baton Rouge are prone to severe congestion. However, roads are beginning to handle the number of vehicles using them after years of stagnation in road upgrades. Baton Rouge Mayor Kip Holden has instituted an extensive upgrade of East Baton Rouge Parish roads known as the Green Light Plan, geared toward improving areas of congestion on the city's surface streets. With its first project completed in October 2008, it has seen numerous others reach completion as of 2015, with several more under construction and still others yet to break ground.
A circumferential loop freeway was proposed for the greater Baton Rouge metro area to help alleviate congestion on the existing through-town routes. The proposed loop would pass through the outlying parishes of Livingston (running alongside property owned and marketed as an industrial development by Al Coburn, a member of President Mike Grimmer's staff), Ascension, West Baton Rouge, and Iberville, as well as northern East Baton Rouge Parish. This proposal has been subject to much contention, particularly by residents living in the outer parishes through which the loop would pass. Other suggestions considered by the community are upgrading Airline Highway (US 61) to freeway standards in the region as well as establishing more links between East Baton Rouge Parish and its neighboring communities.
Commuting
The average one-way commute time in Baton Rouge is 26.5 minutes, slightly less than the US average of 27.1 minutes. Interstates 10, 110 and 12, which feed into the city, are highly traveled and connected by highways and four-lane roads that connect the downtown business area to surrounding parishes.
According to the 2016 American Community Survey, 81.9% of working Baton Rouge residents commuted by driving alone, 8.5% carpooled, 3% used public transportation, and 2.4% walked. About 1.2% used all other forms of transportation, including taxi, bicycle, and motorcycle. About 3.1% worked at home. The city of Baton Rouge has a higher than average percentage of households without a car. In 2015, 10.4 percent of Baton Rouge households lacked a car, and increased slightly to 11.4 percent in 2016. The national average is 8.7 percent in 2016. Baton Rouge averaged 1.55 cars per household in 2016, compared to a national average of 1.8.
Airport
Located 10 minutes north of downtown near Baker, the Baton Rouge Metropolitan Airport connects the area with the four major airline hubs serving the southern United States. Commercial carriers include American Eagle, United Airlines, ViaAir, and Delta Air Lines. Nonstop service is available to Atlanta, Dallas-Ft. Worth, Houston, Austin, Orlando Sanford and Charlotte.
Rail
Three major rail lines, Kansas City Southern, Union Pacific, and Canadian National provide railroad freight service to Baton Rouge. Since 2006, Baton Rouge and New Orleans leaders as well as the state government have been pushing to secure funding for a new high-speed rail passenger line between downtown Baton Rouge and downtown New Orleans, with several stops in between.
Buses and other mass transit
Capital Area Transit System (CATS) provides urban transportation throughout Baton Rouge, including service to Southern University, Baton Rouge Community College, and Louisiana State University. Many CATS buses are equipped with bike racks for commuters to easily combine biking with bus transit.
Greyhound Bus Lines, offering passenger and cargo service throughout the United States, has a downtown terminal on Florida Boulevard.
Sister cities
Cairo, Cairo Governorate, Egypt (since 1951)
Rouen, Seine-Maritime, France (since 1963)
Taichung, Taiwan (since 1976)
Ciudad Obregón, Sonora, Mexico (since 1977)
Port-au-Prince, Ouest, Haiti (since 1978)
Liège, Liège Province, Belgium (since 1985)
Aix-en-Provence, Bouches-du-Rhône, France (since 1987)
Córdoba, Veracruz, Mexico (since 2002)
Heze, Shandong, People's Republic of China (since 2008)
Malatya, Malatya Province, Turkey (since 2009)
Guiyang, Guizhou, People's Republic of China (since 2010)
See also
Baton Rouge Police Department
Cancer Alley
East Baton Rouge Parish Sheriff's Office
Louisiana Technology Park
List of people from Baton Rouge, Louisiana
List of U.S. cities with large Black populations
Explanatory notes
References
Citations
General sources
Frey, Fred Jr. Above Baton Rouge: A Pilot's View Then and Now. Baton Rouge, LA: Louisiana State University Press, 2008. .
External links
City of Baton Rouge official website
Visit Baton Rouge Convention & Visitor's Bureau
The Baton Rouge Digital Archive from the East Baton Rouge Parish Library
Selected Economic Data, Baton Rouge Area, 2012
Geology and geological hazards
Heinrich, P. V., and W. J. Autin, 2000, Baton Rouge 30 × 60 minute geologic quadrangle. Louisiana Geological Survey, Baton Rouge, Louisiana.
McCulloh, R. P., 2001, Active Faults in East Baton Rouge Parish, Louisiana. Public Information Series, no. 8, Louisiana Geological Survey, Baton Rouge, Louisiana.
McCulloh, R. P., 2008a, The Scotlandville, Denham Springs, and Baton Rouge Faults—A Map Guide for Real Estate Buyers, Sellers, and Developers in the Greater Baton Rouge Area. Public Information Series, no. 13, Louisiana Geological Survey, Baton Rouge, Louisiana.
McCulloh, R. P., 2008b, Field Trip Guide to Selected Locations Along the Baton Rouge Fault Trace Spanning the Pleistocene–Holocene Transition in Western East Baton Rouge Parish. Public Information Series, no. 8, Louisiana Geological Survey, Baton Rouge, Louisiana.
1699 establishments in New France
Cities in East Baton Rouge Parish, Louisiana
Cities in Louisiana
Cities in the Baton Rouge metropolitan area
Inland port cities and towns of the United States
Louisiana populated places on the Mississippi River
B
Populated places established in 1699 |
null | null | Army of Northern Virginia | eng_Latn | The Army of Northern Virginia was the primary military force of the Confederate States of America in the Eastern Theater of the American Civil War. It was also the primary command structure of the Department of Northern Virginia. It was most often arrayed against the Union Army of the Potomac.
Origin
The name Army of Northern Virginia referred to its primary area of operation, as did most Confederate States Army names. The Army originated as the Army of the Potomac, which was organized on June 20, 1861, from all operational forces in northern Virginia. On July 20 and July 21, the Army of the Shenandoah and forces from the District of Harpers Ferry were added. Units from the Army of the Northwest were merged into the Army of the Potomac between March 14 and May 17, 1862. The Army of the Potomac was renamed Army of Northern Virginia on March 14. The Army of the Peninsula was merged into it on April 12, 1862.
Robert E. Lee's biographer, Douglas S. Freeman, asserts that the army received its final name from Lee when he issued orders assuming command on June 1, 1862. However, Freeman does admit that Lee corresponded with Joseph E. Johnston, his predecessor in army command, prior to that date and referred to Johnston's command as the Army of Northern Virginia. Part of the confusion results from the fact that Johnston commanded the Department of Northern Virginia (as of October 22, 1861) and the name Army of Northern Virginia can be seen as an informal consequence of its parent department's name. Jefferson Davis and Johnston did not adopt the name, but it is clear that the organization of units as of March 14 was the same organization that Lee received on June 1, and thus it is generally referred to today as the Army of Northern Virginia, even if that is correct only in retrospect.
In addition to Virginians, it included regiments from all over the Confederacy, some from as far away as Georgia, Texas and Arkansas. One of the most well known was the Texas Brigade, made up of the 1st, 4th, and 5th Texas, and the 3rd Arkansas, which distinguished themselves in numerous battles, such as during their fight for the Devil's Den at the Battle of Gettysburg.
Command under Brigadier General P. G. T. Beauregard
The first commander of the Army of Northern Virginia was General P. G. T. Beauregard (under its previous name, the Confederate Army of the Potomac) from June 20 to July 20, 1861. His forces consisted of six brigades, with various militia and artillery from the former Department of Alexandria. During his command, Gen. Beauregard is noted for creating the battle flag of the army, which came to be the primary battle flag for all corps and forces under the Army of Northern Virginia. The flag was designed due to confusion during battle between the Confederate "Stars and Bars" flag and the flag of the United States. Beauregard continued commanding these troops as the new First Corps under Gen. J. E. Johnston as it was joined by the Army of the Shenandoah on July 20, 1861, when command was relinquished to General J. E. Johnston. The following day this army fought its first major engagement in the First Battle of Manassas.
Command under General J. E. Johnston
With the merging of the Army of the Shenandoah, Gen. Joseph E. Johnston took command from July 20, 1861, until May 31, 1862.
Corps organization under Johnston 1861
First Corps – commanded by General P.G.T. Beauregard
Second Corps – commanded by Maj. Gen. G. W. Smith
Wing organization under Johnston 1862
Left Wing – commanded by Maj. Gen. D. H. Hill
Center Wing – commanded by Maj. Gen. James Longstreet
Right Wing – commanded by Maj. Gen. John B. Magruder
Reserve – commanded by Maj. Gen. G. W. Smith
Under the command of Johnston, the Army immediately entered into the First Battle of Manassas. On October 22, 1861, the Department of Northern Virginia was officially created, officially ending the Army of the Potomac. The department comprised three districts: Aquia District, Potomac District, and the Valley District. In April 1862, the department was expanded to include the Departments of Norfolk and the Peninsula (of Virginia). Gen. Johnston was eventually forced into maneuvering the Army southward to the defenses of Richmond during the opening of the Peninsula Campaign, where it conducted delay and defend tactics until Johnston was severely wounded at the Battle of Seven Pines.
During the months after the First Battle of Bull Run, Johnston organized his Shenandoah Army and Beauregard's Potomac Army into two divisions under a unified command with Gustavus Smith and James Longstreet as division commanders. Beauregard quarreled with Johnston and was transferred to the Western theater over the winter months. Jackson was sent to the Shenandoah Valley in October 1861, initially with his own old Stonewall Brigade and later with two other brigades from Western Virginia. Several newly arrived brigades were added to Johnston's army in late 1861-early 1862.
When the Peninsula Campaign began, Johnston took his army down to the Richmond environs where it was merged with several smaller Confederate commands, including a division led by D.H. Hill as well as Benjamin Huger's Department of Norfolk, John Magruder's Army of the Peninsula, and miscellaneous brigades and regiments pulled from various Southern states. Richard Ewell was elevated to division command in the spring of 1862 and sent to join Jackson in the Valley.
On May 27, an additional new division was created and led by A.P. Hill consisting of several new brigades from the Carolinas, Georgia, and Virginia, soon augmented with James Archer's brigade from Smith's division. At Seven Pines, Longstreet and Smith served as temporary wing commanders, and operational control of their divisions went to Brig. Gen William H.C. Whiting and Brig. Gen Richard H. Anderson.
Temporary command under Major General G. W. Smith
Maj. Gen. Gustavus Woodson Smith commanded the ANV on May 31, 1862, following the wounding of Gen. J. E. Johnston during the Battle of Seven Pines. With Smith seemingly having a nervous breakdown, President Jefferson Davis drafted orders to place Gen. Robert E. Lee in command the following day.
Command under General R. E. Lee
On June 1, 1862, its final and most famous leader, General Robert E. Lee, took command after Johnston was wounded and Smith suffered what may have been a nervous breakdown at the Battle of Seven Pines. William Whiting received permanent command of Smith's division, while Richard Anderson reverted to brigade command. Longstreet served as a wing commander for part of the Seven Days Battles and Anderson had operational command of the division at Glendale.
The cavalry, organized into a division on August 17, 1862, and into a corps on
September 9, 1863, was commanded by Maj. Gen. J.E.B. Stuart until May 11, 1864 (the day he was mortally wounded). The cavalry corps was then temporarily split into divisions, but was merged again on August 11, 1864, under command of Lt. Gen. Wade Hampton III. The Reserve Artillery was commanded by Brig. Gen. William N. Pendleton for most of the war.
During the Seven Days Battles, Lee had eleven separate divisions under his command. Aside from the original core army that had been led by Johnston, there were assorted other commands from the Richmond area and North Carolina, as well as Jackson's Valley Army. The inexperience and poor coordination of the army led to the failure of Lee's plans to destroy the Army of the Potomac. As soon as the Seven Days Battles were over, Lee reorganized his army into two corps commanded by Jackson and Longstreet. He removed several generals who had turned in a less-than-inspiring performance in the Seven Days Battles, including John Magruder and Benjamin Huger.
Jackson had five divisions, the commands of A.P. Hill, Ewell, D.H. Hill, and Winder. Longstreet had six divisions commanded by Richard Anderson (formerly Benjamin Huger's division), Cadmus M. Wilcox, James L. Kemper (each commanding half of Longstreet's former division), John Bell Hood (formerly William Whiting's division), David Rumph Jones, and Lafayette McLaws. D.H. Hill's and McLaws's divisions were left behind in the Richmond area and did not participate in the Northern Virginia campaign. The army was also joined for the Northern Virginia and Maryland Campaigns by Nathan G. Evans's independent South Carolina brigade and a North Carolina brigade led by Brig. Gen Thomas Drayton.
During the Maryland Campaign, D.H. Hill rejoined the main army along with Lafayette McLaws. Kemper's division was merged with the division of David R. Jones, a more senior, experienced officer, and Kemper reverted to brigade command. In addition, Robert Ransom commanded two brigades from the Department of North Carolina. At Antietam, Longstreet commanded the divisions of Anderson, McLaws, Jones, Hood, and Ransom while Jackson had the divisions of John R. Jones, Alexander Lawton, A.P. Hill, and D.H. Hill.
The Northern Virginia and Maryland Campaigns still showed numerous defects in the organization and leadership of the Army of Northern Virginia, particularly the high rate of straggling and desertion during the invasion of Maryland. Lee had fewer than 40,000 men on the field at Antietam, the smallest his army would be until the Appomattox Campaign, and the battle was largely fought on autopilot with minimal involvement by the senior officers in the army.
During the Fredericksburg Campaign, Longstreet had the divisions of Anderson, Hood, McLaws, Ransom, and George Pickett, who had just returned to action after months of convalescence from a wound sustained at the Battle of Gaines's Mill. Jackson had the divisions of D.H. Hill, A.P. Hill, Jubal Early, and Elisha Paxton. Robert Ransom's division returned to North Carolina after Fredericksburg. D.H. Hill also departed after quarreling with Lee.
In the Chancellorsville Campaign, Longstreet was sent with Pickett and Hood to the Richmond area. His other two divisions remained with the main army; they were directly commanded by Lee during this time. Robert Rodes took over D.H. Hill's division. Jackson was mortally wounded during the Battle of Chancellorsville. Afterwards, Lee divided the army into three corps with three divisions each. Longstreet got the divisions of Pickett, McLaws, and Hood; A.P. Hill got the divisions of Harry Heth, William D. Pender, and Richard Anderson; and Richard Ewell (returning to action after almost a year of recovering from the loss of a leg at Second Bull Run) got the divisions of Robert Rodes, Jubal Early, and Edward "Allegheny" Johnson. A Fourth Corps, under Lt. Gen. Richard H. Anderson, was organized on October 19, 1864; on April 8, 1865, it was merged into the Second Corps. The commanders of the first three corps changed frequently in 1864 and 1865.
By the time of the Pennsylvania invasion, Lee had fixed the organizational defects that plagued the army during its early campaigns and the straggling problems of the Maryland Campaign did not repeat themselves.
After taking over command in mid-1862, Lee began preparing to lead the Army of Northern Virginia for the first time. However, his aggressiveness to attack the Union led to the loss of many troops especially at the Battle of Antietam, which ended up being a turning point in the war for the Union. After the costly victories during the Seven Days Battles and at Second Manassas in August 1862, Lee had now lost a total of 30,000 of his approximately 92,000 troops within three months of becoming the Confederate's top general. Lee then planned to take his troops north into Maryland to destroy a critical railroad bridge across the Susquehanna River at Harrisburg in a letter written to President Davis. Lee even questioned his own plan, as he wrote, "I am aware that the movement is attended with much risk, yet I do not consider success impossible..." In addition, historians question Lee's aggressiveness to move his army to Maryland. "There can be no sort of doubt that Lee underestimated the exhaustion of his army after Second Manassas. That is, in reality, the major criticism of the Maryland operation: he carried worn-out men across the Potomac." His men were also underarmed and underfed, so the journey to Maryland added to the overall exhaustion. Once Lee arrived in Maryland and was preparing for Antietam, he made another controversial decision. Against the advice from General Longstreet and Jackson, Lee split his troops into four parts to attack the Union from different fronts. Clearly outnumbered and opposed to Lee's plan, Longstreet stated, "General, I wish we could stand still and let the damned Yankees come to us!" As the fighting played out on September 17, 1862, known as the bloodiest single-day battle in American history, the battles at Dunker Church and Burnside's Bridge proved to be too much for Lee and his Confederate army. Luckily for Lee, the arrival of A.P. Hill's troops and the mixture of McClellan's and Burnside's sluggishness, saved Lee's Army of Northern Virginia and allowed them to barely hold off the Union in Maryland.
Corps organization under Lee
Although the Army of Northern Virginia swelled and shrank over time, its units of organization consisted primarily of corps, earlier referred to as "wings" or "commands":
First Corps (Longstreet's Corps)
Second Corps (Jackson's Corps in 1862/1863)
Third Corps (A.P. Hill's Corps)
Fourth Corps (Anderson's Corps)
Cavalry Corps
Campaigns and battles
The Army fought in a number of campaigns and battles, including:
On April 9, 1865, the Army of Northern Virginia surrendered to the Army of the Potomac at Appomattox Court House, effectively ending the Civil War, with General Lee signing the papers of surrender to General Ulysses S. Grant. The day after his surrender, Lee issued his Farewell Address to the Army of Northern Virginia.
Organization of the Army
Department of Northern Virginia, October 22, 1861
The Military Department of Northern Virginia was embattled on October 22, 1861. The department initially consisted of three districts under the overall command of General Joseph E. Johnston.
On February 28, 1862, there were 47,617 soldiers present for duty to the military district. The Cavalry Brigade was provided from the Potomac's Military District and under direct control from the Defense District. The artillery formed an Artillery Corps with 109 cannons.
Organization April 30, 1862
The Army of Northern Virginia was established on March 14, 1862, again under Johnston. Though the military department stayed existent its role changed into an administrative division for most of the war.
At the outset of the Peninsula Campaign the Army of Northern Virginia had more than 55,633 soldiers. The cannon was assigned to the brigades, as well as the Reserve's artillery. Nominally, Jackson's Corps in the Shenandoah Valley, was subordinate to the Army. Since Jackson led his own campaign at the time of the Peninsula Campaign and was not under Lee's direct command this overview does not include his three divisions.
The Army's organization soon proved inept in the course of the Peninsula Campaign. The corps-like structure was rearranged before the Seven Days Battle to converge with the requirements of actual command. In the course of this battle the Army featured two Corps; Jackson's and Magruder's, with four and three divisions respectively, and three actual divisions with five to six brigades. Also the Defense District of North Carolina answered directly to the Army as well as the Reserve Artillery with six battalions and the cavalry with six regiments. The army's complete strength was about 90,000 soldiers. The exact strength cannot be determined, because only a few notes for actual provisionings survived. The estimated strength results, if not explicitly noted, from in-battle dispatches.
Organization at the setout of the Northern Virginia Campaign
The Seven Days Battle showed the Army still suffered from insufficient organization in army command. General Lee subdivided the army again, but this time only with single commands. He introduced a corps-like structure of command, and as an intermediate army management he named the left and right wing. The Army was organized on August 28, 1862, as follows.
The Army's Reserve Artillery consisted of one regiment and two battalions. They stayed in the area of Richmond in the course of the whole Northern Virginia Campaign and only returned on September 3, 1862, to the Army. Major General Hill's Division also remained in the eastern parts of Richmond with the order to bind McClellan's attention as long as possible. As it became predictable that the Army of the Potomac would soon be transferred to support Pope, Lee ordered the Division north. Hill never entered battle in the campaign. A total of about 54,000 soldiers saw action throughout the campaign.
Organization at the beginning of the Maryland Campaign
The Army's losses before and following the Battle of Second Manassas needed to be replaced before the Maryland Campaign could commence. While fundamental changes in the Army's command structure were not necessary, General Lee exchanged divisions and brigades or added additional strength to some. The wings of the Army were now officially called 'Corps'. In the Maryland Campaign the Army was subdivided as follows.
While organization of the corps was found to be generally reliable, the corps' subdivision into four or five divisions hampered overall ease of command. General Lee had already considered before the Battle of Antietam to slim down the overall structure, but intended there be no changes in leadership. The Confederate Congress authorized the establishment of the Corps, and President Davis affirmed the assignment of the commanders and promoted Major Generals Longstreet and Jackson to Lieutenant Generals. General Lee announced this in Special Order 234 on November 6, 1862. About 60,000 soldiers served at the Maryland Campaign.
Fredericksburg and Chancellorsville
In the days following the Battle of Chancellorsville there were no changes in the army's command structure or hierarchy. The army replaced its own losses with new recruits and soldiers returning to duty. Lee made demands that all regiments had to be consolidated solely with recruits originating from their corresponding home states.
Following the Battle of Fredericksburg the Army of Northern Virginia could muster over 72.497 soldiers not counting other personnel. Not figuring into this overall number is the fact that Lee had made the decision to station a whole division and a single cavalry brigade at Shenandoah Valley for protection at this time and these troops are missing from the total estimates.
During the autumn of 1862 lasting throughout the following winter the army faced the Army of the Potomac at Rappahannock. Thereout arose a new problem: Because of the necessity to always show its presence to the enemy there were only limited supplies available for the army from the surrounding villages and towns. Any army at these times supplied itself along the way while moving across the theater of war. Even ordering supplies through the use of the rail, if it was an available option at all, took up considerable time and efforts and supply treks by wagons were potentially prone to enemy assaults. This left Lee with few options and therefore he gave the special order on December 24, 1862, to move half of his artillery into the hinterlands to have his horses better supplied. He forbade to assign all of the horses to any other task than official assignments.
On February 15, 1863, Lee rearranged his artillery. Six battalions were assigned to both corps, and the reserve was composed out of two more battalions. The Confederate War Department strongly suggested in position papers to Lee dating from the February 18th, 1862 to station two divisions at the Atlantic coast. Lee was not averse to any such demands, probably very much on the contrary, because of the tight position he was in with the general lack of provisions, so he gave orders to General Longstreet to have it done so. Both of the assigned divisions only returned to join the Army of Northern Virginia again following the Battle of Chancellorsville. The army was composed at this battle out of over 61.500 soldiers.
Organization from May 30, 1863, until April 9, 1865
Lee took Jackson's death as an opportunity to subdivide the North Virginia Corps again. President Jefferson Davis agreed to the subdivision and ordered Lee in his Special Order Nr. 146 to reorganize the Army.
Lee ordered the artillery battalions of the Reserve Artillery to serve directly with the Corps for the duration of the Gettysburg Campaign. The Army of Northern Virginia now comprised a total of 75,054 soldiers at the Battle of Gettysburg.
The army fielded more than 241 cannons following the Battle of Gettysburg.
On September 9, General Lee had to dispatch the First Corps to Braxton Bragg's Army of Tennessee. Following this the army was resubordinated again. Changes were not significant; only the cavalry saw important reorganization.
The Army's strength was then 55,221 soldiers. The changes in command until December 31, 1863, were only minor. Cooke's Brigade was assigned to serve with Heth's Division, Hampton's Division grew by a cavalry brigade and the Third Corps gained an additional artillery battalion. Imboden's Command remained at Shenandoah Valley and was taken over by Major General Early as the Defense District of Shenandoah Valley. The strength of the army was 54,715 men on December 31.
The organization of the Army of Northern Virginia did not change until the end of the war. The Army featured several corps, the corps featured several divisions, and the artillery was divided between the corps. The strength of the Army grew in the first six months from about 46,380 to 62,230 soldiers. The army was assigned in July to the Defense District of North Carolina and Richmond. In the course of the Richmond-Petersburg Campaign the number of soldiers temporarily grew to 82,633 while parts of the Army were under the command by Lieutenant General Early in Shenandoah Valley.
In 1864 the Army of Northern Virginia fought against the more than twice as strong Potomac-, James- and Shenandoah Army in Grant's Overland Campaign, Early's Raid against the Baltimore and Ohio Railroad, the Richmond-Petersburg Campaign, and Shenandoah Campaign in the Shenandoah Valley. The Army's organization for January 31, 1865 because 69,659 soldiers were fit for battle, but a minimum of 4,500 had no rifles.
Defeat and surrender
Following Lieutenant General A.P. Hill's death on April 2, 1865, the Third Corps was dissolved and assigned to the First Corps. On April 9, 1865, General Lee surrendered. One day later he thanked his men and his officers for their bravery and sturdiness and announced the dismissal of all troops on their word of honor in General Order No. 9. The listings of the Army of Northern Virginia say that 28,231 soldiers were dismissed on their word of honor on April 10, 1865.
See also
Confederate States Army
Flags of the Confederate States of America
Modern display of the Confederate battle flag
Notes
References
Eicher, John H., and Eicher, David J., Civil War High Commands, Stanford University Press, 2001, .
Douglas S. Freeman|Freeman, Douglas S., R. E. Lee, A Biography (4 volumes), Scribners, 1934.
Freeman, Douglas S. R. E. Lee. A Biography. 4. Bde., Charles Scribner's Sons, New York und London 1934f. online here
Freeman, Douglas S. Lee's Lieutenants. A Study in Command. 3 Bde., Scribners, New York 1942–1944.
Katcher, Philip R. N. & Youens, Michael: The Army of Northern Virginia - Osprey Verlag 1975 Men at Arms Series Book Nr. 37 -
Katcher, Philip R. N. & Volstad Ron: American Civil War Armies 1 - Confederate Troops - Osprey Verlag 1986 Men at Arms Series Book Nr. 170 -
Katcher, Philip R. N. & Volstad Ron: American Civil War Armies 3 - Specialist Troops - Osprey Verlag 1987 Men at Arms Series Book Nr. 179 -
Silkenat, David. Raising the White Flag: How Surrender Defined the American Civil War. Chapel Hill: University of North Carolina Press, 2019. .
United States. War Dept.: The War of the Rebellion: a Compilation of the Official Records of the Union and Confederate Armies. Govt. Print. Off., Washington 1880–1901, online here.
Further reading
[http://uapress.ua.edu/product/Small-but-Spartan-Vol.,4736.aspx “A Small but Spartan Vol.: The Florida Brigade in Lee’s Army of Northern Virginia,] by Zack C. Waters and James C. Edmonds, 2010, published by the University of Alabama Press
Young, III, Alfred C. Lee's Army During the Overland Campaign: A Numerical Study'' (Louisiana State University Press; 2013) 400 pages; argues that Lee's Army was far larger in strength and suffered much higher casualties than previously believed.
External links
1861 establishments in the Confederate States of America
1865 disestablishments in the Confederate States of America
Eastern Theater of the American Civil War
Northern Virginia
Military units and formations established in 1861
Military units and formations disestablished in 1865
Northern Virginia
P. G. T. Beauregard
Joseph E. Johnston
Robert E. Lee |
null | null | Lonely Mountain | eng_Latn | In J. R. R. Tolkien's Middle-earth legendarium, the Lonely Mountain is a mountain northeast of Mirkwood. It is the location of the Dwarvish Kingdom under the Mountain. The town of Dale lies in a vale on its southern slopes.
In The Lord of the Rings, the mountain is called by the Sindarin name Erebor. The Lonely Mountain is the goal of the protagonists in The Hobbit, and the scene of the climax.
The Lonely Mountain is a symbol of adventure in The Hobbit, and of the titular Hobbit Bilbo Baggins's maturation as an individual.
Fictional mountain
Erebor stood hundreds of miles from the nearest mountain range. Tolkien's rendering of Thrór's map in The Hobbit shows it with six ridges stretching out from a central peak that was snowcapped well into spring. The whole mountain was perhaps ten miles in diameter; it contained fabulous lodes of gold and jewels.
Origins of the Kingdom under the Mountain
Erebor became the home of the Folk of Durin, a clan of Dwarves known as the Longbeards, after they were driven from their ancestral home of Khazad-dûm. In the latter days of the Third Age, this Kingdom under the Mountain held one of the largest dwarvish treasure hoards in Middle-earth. Dale, a town of Men built between the two southern spurs of Erebor, grew in harmony with the dwarves. The Kingdom under the Mountain was founded by Thráin I the Old, who discovered the Arkenstone there. His son, Thorin I, left the mountain with much of the Folk of Durin to live in the Ered Mithrin (Grey Mountains) on account of the great riches to be found in that range. After dragons plundered their hoards, the Longbeards, led now by Thrór, a descendant of Thorin, returned to Erebor to take up the title King under the Mountain. Under Thrór's reign, Erebor became a great stronghold where the dwarves became numerous and prosperous, well known for the making of matchless weapons and armour.
Erebor in The Hobbit
In the Third Age, while the young Thorin II Oakenshield was out hunting, the dragon Smaug flew south from the Grey Mountains, killed all the dwarves he could find, and destroyed the town of Dale. Smaug then took over the mountain, using the dwarves' hoard as a bed. King Thrór, his son Thráin II, and several companions escaped death by a secret door. Although Thrór and Thráin later perished, Thorin lived in exile in the Ered Luin, far to the west. On a journey, he met the wizard Gandalf. Together they formed a plan to reclaim the mountain. Gandalf insisted that burglary was the best approach and recommended the hobbit Bilbo Baggins, whom he represented to be a professional thief.
Bilbo, Thorin, and Thorin's company of twelve other Dwarves travelled to the Lonely Mountain to regain the treasure. They planned to use the secret door, whose key and map Gandalf had managed to obtain from Thráin, whom he had found at the point of death in the pits of Dol Guldur. On Durin's Day, when the setting sun and the last moon of autumn were in the sky together, the day's last sunlight would fall on the door and expose its keyhole so that it could be unlocked. By a fortunate coincidence, this happened soon after Bilbo and the Dwarves arrived, and the Hobbit was able to enter the mountain and steal a golden cup.
Smaug, enraged by the theft, emerged from the mountain and flew south to destroy Lake-town, which he thought to be the source of the "thieves". During this attack Smaug was killed by Bard the Bowman; Thorin claimed the mountain on learning of Smaug's demise. However, the Men of Esgaroth, supported by Thranduil and the Elves of Mirkwood, marched in force to the mountain to demand a part of the dragon's hoard as recompense for the destruction. Thorin, mad with greed, refused all claims and sent word to his second cousin Dáin II Ironfoot, chief of the Dwarves of the Iron Hills, who brought reinforcements to the aid of Thorin and Company. However before any battle began, an army of Orcs and Wargs descended on Erebor. Dwarves, Elves, and Men joined ranks against them, which led to the Battle of Five Armies. During this battle, Thorin's nephews Fíli and Kíli were killed, and Thorin himself was mortally injured; he died shortly afterwards. The title of King under the Mountain passed to Dáin.
Erebor in The Lord of the Rings
With the restoration of the Kingdom under the Mountain the area became prosperous again. Dale was rebuilt under Bard's leadership, and Dwarves and Men reforged their friendship. Some of the Dwarves, led by Balin, left Erebor to reclaim the ancient Dwarvish Kingdom of Khazad-dûm (also known as Moria). They established a colony there but five years later Balin was killed by an Orc, and soon after Moria was overrun by Orcs and the rest of the Dwarves were killed.
Gimli, a dwarf of Erebor and the son of Glóin, one of Thorin's twelve companions, was chosen to represent his people in the Fellowship of the Ring and helped Aragorn II regain the throne of Gondor.
In the War of the Ring, an emissary from Sauron, the lord of Mordor, twice came to Erebor and spoke to Dáin Ironfoot, who was still King under the Mountain. The messenger asked for assistance in finding Bilbo Baggins and retrieving a stolen ring, and in return offered Moria and three of the seven Dwarf rings to Dáin, but he said neither Yea or Nay. Sauron's northern army, which included many Easterlings, then attacked; Dale was overrun, and many Dwarves and Men took refuge in Erebor, which was promptly surrounded. Dáin was killed before the gates of Erebor defending the body of his fallen ally King Brand of Dale. Dáin's son Thorin III Stonehelm and King Bard II withstood the siege and routed Sauron's forces.
Reception
Dorothy Matthews, viewing The Hobbit as a psychological quest, writes that the Lonely Mountain is an apt symbol of Bilbo's maturation as an individual, as the place where he takes on a leadership role and acts and makes decisions independently. Jared Lobdell comments that he is "profoundly unsympathetic" to Matthews's approach, but that she "carries it off well". Lobdell explains, citing C. S. Lewis's essay "Psychoanalysis and Literary Criticism", that many different stories could, for instance, have the same Freudian interpretation, but be quite different as literature; but that a psychoanalytic approach is at least richer than a purely materialistic one.
William H. Green calls the Lonely Mountain the fourth and final stage of Bilbo's education. He identifies multiple parallels and repetitions of structure between the stages, each one involving a journey, privation, and "unlikely escape". The Lonely Mountain stage, too, symbolically echoes the first stage in the Shire: before setting out, Bilbo was peacefully smoking a pipe of tobacco at his own front door; at the mountain, the smoke is the dragon's, and its meaning is anything but peaceful. The Christian writer Joseph Pearce views the journey to the Lonely Mountain as a "pilgrimage of grace", a Christian bildungsroman, at its deepest level. Pearce states further that Bilbo's quest to the mountain parallel's Frodo's quest to a different mountain, Mount Doom, which he calls "a mirror of Everyman's journey through life".
Tom Shippey notes that in The Hobbit, the lonely mountain is a symbol of adventure, and the "true end" of the story is the moment when Bilbo looks back from a high pass and sees "There far away was the Lonely Mountain on the edge of eyesight. On its highest peak snow yet unmelted was gleaming pale. 'So comes snow after fire, and even dragons have their ending!' said Bilbo, and he turned his back on his adventure."
Amelia Harper, in the J. R. R. Tolkien Encyclopedia, writes that the mountain's history, as usual for the Dwarves, was a tale of "beauty gained and lives lost".
Adaptations
The Lonely Mountain: Lair of Smaug the Dragon is a board game produced in 1985 by Iron Crown Enterprises, designed by Coleman Charlton, which features groups of adventurers, either Dwarves, Elves, Orcs or Men entering Smaug's Lair to capture his treasure before he awakens.
"Erebor", specifically the southern spurs of the Mountain and Dale, is a playable map in The Lord of the Rings: The Battle for Middle-earth II. It has three gates, including the one Tolkien described and two which cannot be closed, to allow those playing as invading forces to easily enter the stronghold.
The Lonely Mountain appears in Peter Jackson's film adaptations of The Hobbit: An Unexpected Journey, The Desolation of Smaug, and The Battle of the Five Armies. The actual setting was Mount Ruapehu in New Zealand.
In astronomy
The International Astronomical Union names all mountains on Saturn's moon Titan after mountains in Tolkien's work. In 2012, they named a mountain on Titan "Erebor Mons" after the Lonely Mountain.
References
Primary
This list identifies each item's location in Tolkien's writings.
Secondary
Middle-earth realms
Fictional kingdoms
Fictional mountains |
null | null | Timeline of rocket and missile technology | eng_Latn | This article gives a concise timeline of rocket and missile technology.
11th century
11th century AD - The first documented record of gunpowder and the fire arrow, an early form of rocketry, appears in the Chinese text Wujing Zongyao.
17th century-19th century
1633 - Lagâri Hasan Çelebi launched a 7-winged rocket using 50 okka (140 lbs) of gunpowder from Sarayburnu, the point below Topkapı Palace in Istanbul.
1650 - Artis Magnae Artilleriae pars prima ("Great Art of Artillery, the First Part") is printed in Amsterdam, about a year before the death of its author, Kazimierz Siemienowicz.
1664 - A "space rocket" is imagined as a future technology to be studied in France and its drawing is ordered by French finance minister Colbert; designed by Le Brun on a Gobelins tapestry (see: French space program)
1798 - Tipu Sultan, the King of the state of Mysore in India, develops and uses iron rockets against the British Army (see Mysorean rockets).
1801 - The British Army develops the Congreve rocket based on weapons used against them by Tipu Sultan.
1806 - Claude Ruggieri, an Italian living in France, launched animals on rockets and recovered them using parachutes. He was prevented from launching a child by police.
1813 - "A Treatise on the Motion of Rockets" by William Moore – first appearance of the rocket equation
1818 - Henry Trengrouse demonstrates his rocket apparatus for projecting a lifeline from a wrecked ship to the shore, later widely adopted
1844 - William Hale invents the spin-stabilized rocket
1861 - William Leitch publishes an essay "A Journey Through Space" (later published in his book God's Glory in the Heavens (1862)) as a humorous science fantasy story about a space gun launching a manned spacecraft equipped with rockets for landing on the Moon, but eventually used for another orbital maneuver.
20th century
1902 - French cinema pioneer Georges Méliès directs A Trip to the Moon, the first film about space travel.
1903 - Konstantin Tsiolkovsky begins a series of papers discussing the use of rocketry to reach outer space, space suits, and colonization of the Solar System. Two key points discussed in his works are liquid fuels and staging.
1913 - Without knowing the work of Russian mathematician Konstantin Tsiolkovsky, French engineer Robert Esnault-Pelterie derived the equations for space flight, produced a paper that presented the rocket equation and calculated the energies required to reach the Moon and nearby planets.
1916 - first use of rockets (with the solid fuel Le Prieur rocket) for both air-to-air attacks, and air to ground.
1922 - Hermann Oberth publishes his scientific work about rocketry and space exploration: Die Rakete zu den Planetenräumen ("By Rocket into Planetary Space").
1924 - Society for Studies of Interplanetary Travel founded in Moscow by Konstantin Tsiolkovsky, Friedrich Zander and 200 other space and rocket experts
1926 - Robert Goddard launches the first liquid fuel rocket. This is considered by some to be the start of the Space Age.
1927 - Verein für Raumschiffahrt (VfR - "Spaceflight Society") founded in Germany.
1928 - World's first large-scale rocket program, Opel RAK initiated by Max Valier and Fritz von Opel, achieving speed records for ground and rail vehicles in 1928
1929 - Opel RAK head of rocket technology Friedrich Wilhelm Sander launches two liquid-fuel rockets on April 10 and 12, 1929 as of 21 cm in diameter and with a length of 74 cm, weighing 7 kg empty and 16 kg with fuel, maximum thrust was 45 to 50 kp, total burning time 132 seconds.
1929 - World's first public flight of a manned rocket plane, Opel RAK.1, piloted by Fritz von Opel on September 30, 1929
1929 - Woman in the Moon, considered to be one of the first "serious" science fiction films.
1931 - Friedrich Schmiedl attempts the first rocket mail service in Austria
1933 - Sergei Korolev and Mikhail Tikhonravov launch the first liquid-fueled rocket in the Soviet Union
1935 - Emilio Herrera Linares from Spain designed and made the first full-pressured astronaut suit, called the escafandra estratonáutica. The Russians then used a model of Herrera's suit when first flying into space of which the Americans would then later adopt when creating their own space program
1936 - Research on rockets begins at the Guggenheim Aeronautical Laboratory at the California Institute of Technology (GALCIT), the predecessor to the Jet Propulsion Laboratory, under the direction of Frank Malina and Theodore von Kármán
1937 - Peenemünde Army Research Center founded in Germany
1938 - The Projectile Development Establishment founded at Fort Halstead for the United Kingdom's research into military solid-fuel rockets.
1939 - Katyusha multiple rocket launchers () are a type of rocket artillery first built and fielded by the Soviet Union.
1941 - French rocket EA-41 is launched, being the first European liquid propellant working rocket (It was, however, preceded by the Peenemunde A5 and Soviet experiments.)
1941 - Jet Assisted Take Off JATO installed on US Army Air Corp Ercoupe aircraft occurred on 12 August in March Field, California.
1942 - Wernher von Braun and Walter Dornberger launch the first V-2 rocket at Peenemünde in northern Germany.
1942 - A V-2 rocket reaches an altitude of 85 km.
1944 - The V-2 rocket MW 18014 reaches an altitude of 176 km, becoming the first man-made object in space.
1945 - Lothar Sieber dies after the first vertical take-off manned rocket flight in a Bachem Ba 349 "Natter"
1945 - Operation Paperclip takes 1,600 German rocket scientists and technicians to the United States
1945 - Operation Osoaviakhim takes 2,000 German rocket scientists and technicians to the Soviet Union
1946 - First flight of the Nike missile, later the first operational surface-to-air guided missile
1947 - The first animals sent into space were fruit flies aboard a V-2 rocket launched from New Mexico, USA
1947 - Chuck Yeager achieves the first manned supersonic flight in a Bell X-1 rocket-powered aircraft
1949 - Willy Ley publishes The Conquest of Space
1952 - 22 May, French Véronique 1 rocket is launched from the Algerian desert.
1952 - Wernher von Braun discusses the technical details of a manned exploration of Mars in Das Marsprojekt.
1953 - Colliers magazine publishes a series of articles on man's future in space, igniting the interest of people around the world. The series includes numerous articles by Ley and von Braun, illustrated by Chesley Bonestell.
1956 - First launch of PGM-17 Thor, the first US ballistic missile and forerunner of the Delta space launch rockets
1957 - Launch of the first ICBM, the USSR's R-7 (8K71), known to NATO as the SS-6 Sapwood.
1957 - The USSR launches Sputnik 1, the first artificial satellite.
1958 - The U.S. launches Explorer 1, the first American artificial satellite, on a Jupiter-C rocket.
1958 - US launches their first ICBM, the Atlas-B (the Atlas-A was a test article only).
1961 - the USSR launches Vostok 1, Yuri Gagarin reached a height of 327 km above Earth and was the first man to orbit Earth.
1961 - US, a Mercury capsule named Freedom 7 with Alan B. Shepard, spacecraft was launched by a Redstone rocket on a ballistic trajectory suborbital flight. It was the first human space mission that landed with pilot still in spacecraft, thus the first complete human spaceflight by then FAI definitions.
1962 - The US launches Mercury MA-6 (Friendship 7) on an Atlas D booster, John Glenn puts America in orbit.
1962 - Pakistan launched Rehbar-I and was the first country in Islamic world to successfully launch a vessel in outer space.
1963 - The USSR launches Vostok 6, Valentina Tereshkova was the first woman (and first civilian) in space and to orbit Earth. She remained in space for nearly three days and orbited the Earth 48 times.
1963 - US X-15 rocket-plane, the first reusable manned spacecraft (suborbital) reaches space, pioneering reusability, carried launch and glide landings.
1965 - USSR Proton rocket, highly successful launch vehicle with notable payloads, Salyut 6 and Salyut 7, Mir, and ISS components
1965 - Robert Salked investigates various single stage to orbit spaceplane concepts
1965 - FR Diamant, first French and European rocket to reach the orbit, France became the third space nation.
1966 - USSR Luna 9, the first soft landing on the Moon
1966 - USSR launches Soyuz spacecraft, longest-running series of spacecraft, eventually serving Soviet, Russian and International space missions.
1968 - USSR Zond 5, two tortoises and smaller biological Earthlings circle the Moon and return safely to Earth.
1968 - US Apollo 8, the first men to reach and orbit the Moon.
1969 - US Apollo 11, first men on the Moon, first lunar surface extravehicular activity.
1975 - EU ESA, creation of the European Space Agency.
1979 - EU Ariane 1, first Ariane European rocket.
1980 - EU Arianespace, creation of Arianespace, world first commercial space transportation company.
1981 - US Space Shuttle pioneers reusability and glide landings
1988 - EU Ariane 4, first launch of the Ariane 4 rocket.
1996 - EU Ariane 5, first flight of the Ariane 5 rocket, self-destructed in flight. After that, Ariane 5 will be the main European rocket for decades.
1998 - US Deep Space 1 is first deep space mission to use an ion thruster for propulsion.
1998 - Russia launch Zarya module which is the first part of the International Space Station.
21st century
2001 - Russian Soyuz spacecraft sent the first space tourist Dennis Tito to International Space Station.
2004 - US-based, first privately developed, manned (suborbital) spaceflight, SpaceShipOne demonstrates reusability.
2008 - SpaceX—with their Falcon 1 rocket—became the first private entity to successfully launch a rocket into orbit.
2012 - The SpaceX Dragon space capsule—launched aboard a Falcon 9 launch vehicle—was the first private spacecraft to successfully dock with another spacecraft, and was also the first private capsule to dock at the International Space Station.
2014 - First booster rocket returning from an orbital trajectory to achieve a zero-velocity-at-zero-altitude propulsive vertical landing. The first-stage booster of Falcon 9 Flight 9 made the first successful controlled ocean soft touchdown of a liquid-rocket-engine orbital booster on April 18, 2014.
2015 - SpaceX's Falcon 9 Flight 20 was the first time that the first stage of an orbital rocket made a successful return and vertical landing.
2017 - SpaceX's Falcon 9 SES-10 was the first time a used orbital rocket made a successful return
2018 - The Electron rocket was the first New-Zealand rocket to achieve orbit. The rocket is also unique in using an electric pump-fed engine. The rocket also carried an additional satellite payload called "Humanity Star", a 1-meter-wide (3 ft) carbon fiber sphere made up of 65 panels that reflect the Sun's light.
2020 - SpaceX's Dragon capsule, rode on the Falcon 9 rocket whilst carrying crew aboard, marking the first time astronauts have been sent to space via a private company.
See also
History of rockets
List of missiles
Lists of rockets
Timeline of heat engine technology
References
Rockets
Rocket
Missile
Rockets and missiles
Missiles and rocksts |
null | null | Timeline of knowledge about galaxies, clusters of galaxies, and large-scale structure | eng_Latn | Timeline of galaxies, clusters of galaxies, and large-scale structure of the universe
Pre-20th century
5th century BC — Democritus proposes that the bright band in the night sky known as the Milky Way might consist of stars,
4th century BC — Aristotle believes the Milky Way to be caused by "the ignition of the fiery exhalation of some stars which were large, numerous and close together" and that the "ignition takes place in the upper part of the atmosphere, in the region of the world which is continuous with the heavenly motions",
964 — Abd al-Rahman al-Sufi (Azophi), a Persian astronomer, makes the first recorded observations of the Andromeda Galaxy and the Large Magellanic Cloud in his Book of Fixed Stars, and which are the first galaxies other than the Milky Way to be observed from Earth,
11th century — Abū Rayḥān al-Bīrūnī, another Persian astronomer, describes the Milky Way galaxy as a collection of numerous nebulous stars,
11th century — Alhazen (Ibn al-Haytham), an Arabian astronomer, refutes Aristotle's theory on the Milky Way by making the first attempt at observing and measuring the Milky Way's parallax, and he thus "determined that because the Milky Way had no parallax, it was very remote from the Earth and did not belong to the atmosphere",
12th century — Avempace (Ibn Bajjah) of Islamic Spain proposes the Milky Way to be made up of many stars but that it appears to be a continuous image due to the effect of refraction in the Earth's atmosphere,
14th century — Ibn Qayyim Al-Jawziyya of Syria proposes the Milky Way galaxy to be "a myriad of tiny stars packed together in the sphere of the fixed stars" and that these stars are larger than planets,
1521 — Ferdinand Magellan observes the Magellanic Clouds during his circumnavigating expedition,
1610 — Galileo Galilei uses a telescope to determine that the bright band on the sky, the "Milky Way", is composed of many faint stars
1612 - Simon Marius using a moderate telescope observes Andromeda and describes as a “flame seen through horn”.
1750 — Thomas Wright discusses galaxies and the flattened shape of the Milky Way and speculates nebulae as separate
1755 — Immanuel Kant drawing on Wright's work conjectures our galaxy is a rotating disk of stars held together by gravity, and that the nebulae are separate such galaxies; he calls them Island Universes
1774 — Charles Messier releases a preliminary list of 45 Messier objects, three of which turn out to be the galaxies including Andromeda and Triangulum. By 1781 the final published list grows to 103 objects, 34 of which turn out to be galaxies.
1785 — William Herschel carried the first attempt to describe the shape of the Milky Way and the position of the Sun in it by carefully counting the number of stars in different regions of the sky. He produced a diagram of the shape of the galaxy with the solar system close to the center.
1845 — Lord Rosse discovers a nebula with a distinct spiral shape
Early 20th century
1912 — Vesto Slipher spectrographic studies of spiral nebulae find high Doppler shifts indicating recessional velocity.
1917 — Heber Curtis find novae in Andromeda Nebula M31 were ten magnitudes fainter than normal giving a distance estimate of 150,000 parsecs supporting the "island universes" or independent galaxies hypothesis for spiral nebulae.
1918 — Harlow Shapley demonstrates that globular clusters are arranged in a spheroid or halo whose center is not the Earth, and hypothesizes, correctly, that its center is the Galactic Center of the galaxy,
26 April 1920 — Harlow Shapley and Heber Curtis debate whether Andromeda Nebula is within the Milky Way. Curtis notes dark lanes in Andromeda resembling the dust clouds in the Milky Way, as well as significant Doppler shift.
1922 — Ernst Öpik distance determination supports Andromeda as extra-galactic object.
1923 — Edwin Hubble resolves the Shapley–Curtis debate by finding Cepheids in the Andromeda Galaxy, definitively proving that there are other galaxies beyond the Milky Way.
1930 — Robert Trumpler uses open cluster observations to quantify the absorption of light by interstellar dust in the galactic plane; this absorption had plagued earlier models of the Milky Way,
1932 — Karl Guthe Jansky discovers radio noise from the center of the Milky Way,
1933 — Fritz Zwicky applies the virial theorem to the Coma Cluster and obtains evidence for unseen mass,
1936 — Edwin Hubble introduces the spiral, barred spiral, elliptical, and irregular galaxy classifications,
1939 — Grote Reber discovers the radio source Cygnus A,
1943 — Carl Keenan Seyfert identifies six spiral galaxies with unusually broad emission lines, named Seyfert galaxies,
1949 — J. G. Bolton, G. J. Stanley, and O. B. Slee identify NGC 4486 (M87) and NGC 5128 as extragalactic radio sources,
Mid-20th century
1953 — Gérard de Vaucouleurs discovers that the galaxies within approximately 200 million light-years of the Virgo Cluster are confined to a giant supercluster disk,
1954 — Walter Baade and Rudolph Minkowski identify the extragalactic optical counterpart of the radio source Cygnus A,
1959 — Hundreds of radio sources are detected by the Cambridge Interferometer which produces the 3C catalogue. Many of these are later found to be distant quasars and radio galaxies
1960 — Thomas Matthews determines the radio position of the 3C source 3C 48 to within 5",
1960 — Allan Sandage optically studies 3C 48 and observes an unusual blue quasistellar object,
1962 — Cyril Hazard, M. B. Mackey, and A. J. Shimmins use lunar occultations to determine a precise position for the quasar 3C 273 and deduce that it is a double source,
1962 — Olin Eggen, Donald Lynden-Bell, and Allan Sandage theorize galaxy formation by a single (relatively) rapid monolithic collapse, with the halo forming first, followed by the disk.
1963 — Maarten Schmidt identifies the redshifted Balmer lines from the quasar 3C 273
1973 — Jeremiah Ostriker and James Peebles discover that the amount of visible matter in the disks of typical spiral galaxies is not enough for Newtonian gravitation to keep the disks from flying apart or drastically changing shape,
1973 — Donald Gudehus finds that the diameters of the brightest cluster galaxies have increased due to merging, the diameters of the faintest cluster galaxies have decreased due to tidal distention, and that the Virgo cluster has a substantial peculiar velocity,
1974 — B. L. Fanaroff and J. M. Riley distinguish between edge-darkened (FR I) and edge-brightened (FR II) radio sources,
1976 — Sandra Faber and Robert Jackson discover the Faber-Jackson relation between the luminosity of an elliptical galaxy and the velocity dispersion in its center. In 1991 the relation is revised by Donald Gudehus,
1977 — R. Brent Tully and Richard Fisher publish the Tully–Fisher relation between the luminosity of an isolated spiral galaxy and the velocity of the flat part of its rotation curve,
1978 — Steve Gregory and Laird Thompson describe the Coma supercluster,
1978 — Donald Gudehus finds evidence that clusters of galaxies are moving at several hundred kilometers per second relative to the cosmic microwave background radiation,
1978 — Vera Rubin, Kent Ford, N. Thonnard, and Albert Bosma measure the rotation curves of several spiral galaxies and find significant deviations from what is predicted by the Newtonian gravitation of visible stars,
1978 — Leonard Searle and Robert Zinn theorize that galaxy formation occurs through the merger of smaller groups.
Late 20th century
1981 — Robert Kirshner, August Oemler, Paul Schechter, and Stephen Shectman find evidence for a giant void in Boötes with a diameter of approximately 100 million light years,
1985 — Robert Antonucci and J. Miller discover that the Seyfert II galaxy NGC 1068 has broad lines which can only be seen in polarized reflected light,
1986 — Amos Yahil, David Walker, and Michael Rowan-Robinson find that the direction of the IRAS galaxy density dipole agrees with the direction of the cosmic microwave background temperature dipole,
1987 — David Burstein, Roger Davies, Alan Dressler, Sandra Faber, Donald Lynden-Bell, R. J. Terlevich, and Gary Wegner claim that a large group of galaxies within about 200 million light years of the Milky Way are moving together towards the "Great Attractor" in the direction of Hydra and Centaurus,
1987 — R. Brent Tully discovers the Pisces–Cetus Supercluster Complex, a structure one billion light years long and 150 million light years wide,
1989 — Margaret Geller and John Huchra discover the "Great Wall", a sheet of galaxies more than 500 million light years long and 200 million wide, but only 15 million light years thick,
1990 — Michael Rowan-Robinson and Tom Broadhurst discover that the IRAS galaxy IRAS F10214+4724 is the brightest known object in the Universe,
1991 — Donald Gudehus discovers a serious systematic bias in certain cluster galaxy data (surface brightness vs. radius parameter, and the method) which affect galaxy distances and evolutionary history; he devises a new distance indicator, the reduced galaxian radius parameter, , which is free of biases,
1992 — First detection of large-scale structure in the Cosmic microwave background indicating the seeds of the first clusters of galaxies in the early Universe
1995 — First detection of small-scale structure in the cosmic microwave background
1995 — Hubble Deep Field survey of galaxies in field 144 arc seconds across.
1998 — The 2dF Galaxy Redshift Survey maps the large-scale structure in a section of the Universe close to the Milky Way
1998 — Hubble Deep Field South
1998 — Discovery of accelerating universe
2000 — Data from several cosmic microwave background experiments give strong evidence that the Universe is "flat" (space is not curved, although space-time is), with important implications for the formation of large-scale structure
Early 21st century
2001 — First data release from the ongoing Sloan Digital Sky Survey
2004 — The European Southern Observatory discovers Abell 1835 IR1916, the most distant galaxy yet seen from Earth.
2004 — The Arcminute Microkelvin Imager begins to map the distribution of distant clusters of galaxies
2005 — Spitzer Space Telescope data confirm what had been considered likely since the early 1990s from radio telescope data, i.e., that the Milky Way Galaxy is a barred spiral galaxy.
2012 — Astronomers report the discovery of the most distant dwarf galaxy yet found, approximately 10 billion light-years away.
2012 — The Huge-LQG, a large quasar group, one of the largest known structures in the universe, is discovered.
2013 — The galaxy Z8 GND 5296 is confirmed by spectroscopy to be one of the most distant galaxies found up to this time. Formed just 700 million years after the Big Bang, expansion of the universe has carried it to its current location, about 13 billion light years away from Earth (30 billion light years comoving distance).
2013 — The Hercules–Corona Borealis Great Wall, a massive galaxy filament and the largest known structure in the universe, was discovered through gamma-ray burst mapping.
2014 — The Laniakea Supercluster, the galaxy supercluster that is home to the Milky Way is defined via a new way of defining superclusters according to the relative velocities of galaxies. The new definition of the local supercluster subsumes the prior defined local supercluster, the Virgo Supercluster, as an appendage.
2020 — Astronomers report the discovery of a large cavity in the Ophiuchus Supercluster, first detected in 2016 and originating from a supermassive black hole with the mass of 10 million solar masses. The cavity is a result of the largest known explosion in the Universe. The formerly active galactic nucleus created it by emitting radiation and particle jets, possibly as a result of a spike in supply of gas to the black hole that could have occurred if a galaxy fell into the centre of the cavity.
2020 — Astronomers report to have discovered the disk galaxy Wolfe Disk, dating back to when the universe was only 1.5 billion years old, possibly indicating the need to revise theories of galaxy formation and evolution.
2020 — The South Pole Wall is a massive cosmic structure formed by a giant wall of galaxies (a galaxy filament) that extends across at least 1.37 billion light-years of space, and is located approximately a half billion light-years away.
2020 — After a 20-year-long survey, astrophysicists of the Sloan Digital Sky Survey publish the largest, most detailed 3D map of the universe so far, fill a gap of 11 billion years in its expansion history, and provide data which supports the theory of a flat geometry of the universe and confirms that different regions seem to be expanding at different speeds.
See also
Illustris project
Large-scale structure of the cosmos
Timeline of astronomical maps, catalogs, and surveys
Timeline of cosmological theories
UniverseMachine
List of largest cosmic structures
References
Galaxies, clusters of galaxies, and large-scale structure, timeline of knowledge about |
null | null | Colon (punctuation) | eng_Latn | The colon is a punctuation mark consisting of two equally sized dots placed on the same vertical line. A colon often precedes an explanation, a list, or a quoted sentence. It is also used between hours and minutes in time, between certain elements in medical journal citations, chapter and verse in Bible citations, and, in the US, for salutations in business letters and other formal letter writing.
History
The English word "colon" is from Latin ( ), itself from Ancient Greek (), meaning "limb", "member", or "portion". In Greek rhetoric and prosody, the term did not refer to punctuation but to the expression or passage itself. A "colon" was a section of a complete thought or passage. From this usage, in palaeography, a colon is a clause or group of clauses written as a line in a manuscript.
In the punctuation system devised by Aristophanes of Byzantium in the 3rd century BC, the end of such a clause was thought to occasion a medium-length breath and was marked by a middot . (This was only intermittently used, but eventually revived as the ano teleia, the modern Greek semicolon.) A double dot symbol , meanwhile, later came to be used as a full stop or to mark a change of speaker. A variant was introduced to English orthography around 1600, marking a pause intermediate between a comma and a full stop. As late as the 18th century, the appropriateness of a colon was still being related to the length of the pause taken when reading the text aloud, but silent reading eventually replaced this with other considerations.
In British English, it was once common for a colon to be followed by a hyphen or dash to indicate a restful pause, in a typographical construction known as the "dog's bollocks", though this usage is now discouraged.
Usage in English
In modern English usage, a complete sentence precedes a colon, while a list, description, explanation, or definition follows it. The elements which follow the colon may or may not be a complete sentence: since the colon is preceded by a sentence, it is a complete sentence whether what follows the colon is another sentence or not. While it is acceptable to capitalise the first letter after the colon in American English, it is not the case in British English, except where a proper noun immediately follows a colon.
Colon used before list
Daequan was so hungry that he ate everything in the house: chips, cold pizza, pretzels and dip, hot dogs, peanut butter, and candy.
Colon used before a description
Bertha is so desperate that she'll date anyone, even William: he's uglier than a squashed toad on the highway, and that's on his good days.
Colon before definition
For years while I was reading Shakespeare's Othello and criticism on it, I had to constantly look up the word "egregious" since the villain uses that word: outstandingly bad or shocking.
Colon before explanation
I guess I can say I had a rough weekend: I had chest pain and spent all Saturday and Sunday in the emergency room.
Some writers use fragments (incomplete sentences) before a colon for emphasis or stylistic preferences (to show a character's voice in literature), as in this example:
Dinner: chips and juice. What a well-rounded diet I have.
The Bedford Handbook describes several uses of a colon. For example, one can use a colon after an independent clause to direct attention to a list, an appositive or a quotation, and it can be used between independent clauses if the second summarizes or explains the first. In non-literary or non-expository uses, one may use a colon after the salutation in a formal letter, to indicate hours and minutes, to show proportions, between a title and subtitle, and between city and publisher in bibliographic entries.
Luca Serianni, an Italian scholar who helped to define and develop the colon as a punctuation mark, identified four punctuational modes for it: syntactical-deductive, syntactical-descriptive, appositive, and segmental.
Syntactical-deductive
The colon introduces the logical consequence, or effect, of a fact stated before.
There was only one possible explanation: the train had never arrived.
Syntactical-descriptive
In this sense the colon introduces a description; in particular, it makes explicit the elements of a set.
I have three sisters: Daphne, Rose, and Suzanne.
Syntactical-descriptive colons may separate the numbers indicating hours, minutes, and seconds in abbreviated measures of time.
The concert begins at 21:45.
The rocket launched at 09:15:05.
British English, however, more frequently uses a point for this purpose:
The programme will begin at 8.00 pm.
You will need to arrive by 14.30.
A colon is also used in the descriptive location of a book verse if the book is divided into verses, such as in the Bible or the Quran:
"Isaiah 42:8"
"Deuteronomy 32:39"
"Quran 10:5"
Appositive
Luruns could not speak: he was drunk.
An appositive colon also separates the subtitle of a work from its principal title. (In effect, the example given above illustrates an appositive use of the colon as an abbreviation for the conjunction 'because'.) Dillon has noted the impact of colons on scholarly articles, but the reliability of colons as a predictor of quality or impact has also been challenged. In titles, neither needs to be a complete sentence as titles do not represent expository writing:
Star Wars Episode VI: Return of the Jedi
Segmental
Like a dash or quotation mark, a segmental colon introduces speech. The segmental function was once a common means of indicating an unmarked quotation on the same line. The following example is from the grammar book The King's English:
Benjamin Franklin proclaimed the virtue of frugality: A penny saved is a penny earned.
This form is still used in written dialogues, such as in a play. The colon indicates that the words following an individual's name are spoken by that individual.
Patient: Doctor, I feel like a pair of curtains.
Doctor: Pull yourself together!
Use of capitals
Use of capitalization or lower-case after a colon varies. In British English, and in most Commonwealth countries, the word following the colon is in lower case unless it is normally capitalized for some other reason, as with proper nouns and acronyms. British English also capitalizes a new sentence introduced by colon's segmental use.
American English goes further and permits writers to similarly capitalize the first word of any independent clause following a colon. This follows the guidelines of some modern American style guides, including those published by the Associated Press and the Modern Language Association. The Chicago Manual of Style, however, requires capitalization only when the colon introduces a direct quotation, a direct question, or two or more complete sentences.
In many European languages, the colon is usually followed by a lower-case letter unless the upper case is required for other reasons, as with British English. German usage requires capitalization of independent clauses following a colon. Dutch further capitalizes the first word of any quotation following a colon, even if it is not a complete sentence on its own.
Spacing
In print, a thin space was traditionally placed before a colon and a thick space after it. In modern English-language printing, no space is placed before a colon and a single space is placed after it. In French-language typing and printing, the traditional rules are preserved.
One or two spaces may be and have been used after a colon. The older convention (designed to be used by monospaced fonts) was to use two spaces after a colon.
Usage in other languages
Suffix separator
In Finnish and Swedish, the colon can appear inside words in a manner similar to the apostrophe in the English possessive case, connecting a grammatical suffix to an abbreviation or initialism, a special symbol, or a digit (e.g., Finnish USA:n and Swedish USA:s for the genitive case of "USA", Finnish %:ssa for the inessive case of "%", or Finnish 20:een for the illative case of "20").
Abbreviation mark
Written Swedish uses colons in contractions, such as S:t for Sankt (Swedish for "Saint") – for example in the name of the Stockholm metro station S:t Eriksplan, and k:a for kyrka ("church") – for instance Svenska k:a (Svenska kyrkan), the Evangelical Lutheran national Church of Sweden. This can even occur in people's names, for example Antonia Ax:son Johnson (Ax:son for Axelson). Early Modern English texts also used colons to mark abbreviations.
Word separator
In Ethiopia, both Amharic and Ge'ez script used and sometimes still use the colon as word separator.
End of sentence or verse
In Armenian, a colon indicates the end of a sentence, similar to a Latin full stop or period.
In Hebrew, the sof pasuq is used in some writings such as prayer books to signal the end of a verse.
Score divider
In German, a colon divides the scores of opponents in sports and games. A result of 149–0 would be written as 149 : 0 in German.
Mathematics and logic
The colon is used in mathematics, cartography, model building, and other fields—in this context it denotes a ratio or a scale, as in 3:1 (pronounced "three to one"). When a ratio is reduced to a simpler form, such as 10:15 to 2:3, this may be expressed with a double colon as 10:15::2:3; this would be read "10 is to 15 as 2 is to 3". This form is also used in tests of logic where the question of "Dog is to Puppy as Cat is to _?" can be expressed as "Dog:Puppy::Cat:_".
In some languages (e.g. German, Russian and French), the colon is the commonly used sign for division (instead of ÷).
The notation | : | may also denote the index of a subgroup.
The notation indicates that is a function with domain and codomain .
The combination with an equal sign () is used for definitions.
In mathematical logic, when using set-builder notation for describing the characterizing property of a set, it is used as an alternative to a vertical bar (which is the ISO 31-11 standard), to mean "such that". Example:
(S is the set of all in (the real numbers) such that is strictly greater than 1 and strictly smaller than 3)
In older literature on mathematical logic, it is used to indicate how expressions should be bracketed (see Glossary of Principia Mathematica).
In type theory and programming language theory, the colon sign after a term is used to indicate its type, sometimes as a replacement to the "∈" symbol. Example:
.
A colon is also sometimes used to indicate a tensor contraction involving two indices, and a double colon (::) for a contraction over four indices.
A colon is also used to denote a parallel sum operation involving two operands (many authors, however, instead use a ∥ sign and a few even a ∗ for this purpose).
Computing
The character was on early typewriters and therefore appeared in most text encodings, such as Baudot code and EBCDIC. It was placed at code 58 in ASCII and from there inherited into Unicode. Unicode also defines several related characters.
, used in IPA.
, used in IPA.
, used by Uralic Phonetic Alphabet.
, compatible with right-to-left text.
, for mathematical usage.
, for use in pretty-printing programming languages.
, see colon (letter), sometimes used in Windows filenames as it is identical to the colon in the Segoe UI font used for filenames. The colon itself is not permitted as it is a reserved character.
, compatibility character for the Chinese Standard GB 18030.
, for compatibility with half/full width fonts.
, compatibility character for the Chinese National Standard CNS 11643.
Programming languages
A number of programming languages, most notably Pascal and Ada, use a colon immediately followed by an equals sign (:=) as the assignment operator, to distinguish it from a single equals which is an equality test (C instead used a single equals as assignment, and a double equals as the equality test).
Many languages including C, C++ and DOS batch files use the colon to indicate the text before it is a label, such as a target for a goto or an introduction to a case in a switch statement. In a related use, Python uses a colon to separate a control statement from the block of statements it controls:
if test(x):
print("test(x) is true!")
else:
print("test(x) is not true...")
In a number of languages, including JavaScript and Python, colons are used to define name-value pairs in a dictionary or object. This is also used by data formats such as JSON. Some other languages use an equals sign.
var obj = {
name: "Charles",
age: 18,
}
The colon is used as part of the ?: conditional operator in C and many other languages.
C++ uses a double colon as the scope resolution operator, and class member access. Most other languages use a period but C++ had to use this for compatibility with C.
Erlang uses a single colon for scope resolution.
In BASIC, it is used as a separator between the statements or instructions in a single line. Most other languages use a semicolon, but BASIC had used semicolon to separate items in print statements.
In Forth, a colon precedes definition of a new word.
Haskell uses a colon (pronounced as "cons", short for "construct") as an operator to add an element to the front of a list:
"child" : ["woman", "man"] -- equals ["child","woman","man"]
while a double colon :: is read as "has type of" (compare scope resolution operator):
("text", False) :: ([Char], Bool)
The ML languages (including Standard ML and OCaml) have the above reversed, where the double colon (::) is used to add an element to the front of a list; and the single colon (:) is used for type guards.
MATLAB uses the colon as a binary operator that generates vectors, as well as to select particular portions of existing matrices.
APL uses the colon
to introduce a control structure element. In this usage it must be the first non-blank character of the line.
after a label name that will be the target of a :goto or a right-pointing arrow (Note: this style of programming is deprecated and programmers are encouraged to use control structures instead).
to separate a guard (boolean expression) from its expression in a dynamic function. Two colons are used for an Error guard (one or more error numbers).
Colon + space are used in class definitions to indicate inheritance.
⍠ (a colon in a box) is used by APL for its variant operator.
The colon is also used in many operating systems commands.
In the esoteric programming language INTERCAL, the colon is called "two-spot" and is used to identify a 32-bit variable—distinct from a spot (.) which identifies a 16-bit variable.
Addresses
Internet URLs use the colon to separate the protocol (such as ) from the hostname or IP address.
In an IPv6 address colons (and one optional double colon) separate up to 8 groups of 16 bits in hexadecimal representation. In a URL a colon follows the initial scheme name (such as HTTP and FTP), and separates a port number from the hostname or IP address.
In Microsoft Windows filenames, the colon is reserved for use in alternate data streams and cannot appear in a filename. It was used as the directory separator in Classic Mac OS, and was difficult to use in early versions of the newer BSD-based macOS due to code swapping the slash and colon to try to preserve this usage. In most systems it is often difficult to put a colon in a filename as the shell interprets it for other purposes.
CP/M and early versions of MSDOS required the colon after the names of devices, such as though this gradually disappeared except for disks (where it had to be between the disk name and the required path representation of the file as in C:\Windows\). This then migrated to use in URLs.
Text markup
It is often used as a single post-fix delimiter, signifying a token keyword had immediately preceded it or the transition from one mode of character string interpretation to another related mode. Some applications, such as the widely used MediaWiki, utilize the colon as both a pre-fix and post-fix delimiter.
In wiki markup, the colon is often used to indent text. Common usage includes separating or marking comments in a discussion as replies, or to distinguish certain parts of a text.
In human-readable text messages, a colon, or multiple colons, is sometimes used to denote an action (similar to how asterisks are used) or to emote (for example, in vBulletin). In the action denotation usage it has the inverse function of quotation marks, denoting actions where unmarked text is assumed to be dialogue. For example:
Tom: Pluto is so small; it should not be considered a planet. It is tiny!
Mark: Oh really? ::drops Pluto on Tom's head:: Still think it's small now?
Colons may also be used for sounds, e.g., ::click::, though sounds can also be denoted by asterisks or other punctuation marks.
Colons can also be used to represent eyes in emoticons.
See also
Semicolon
References
External links
Walden University Guides Punctuation: Colons
Punctuation
Typographical symbols |
null | null | Mission San Diego de Alcalá | eng_Latn | Mission Basilica San Diego de Alcalá () was the first Franciscan mission in The Californias, a province of New Spain. Located in present-day San Diego, California, it was founded on July 16, 1769, by Spanish friar Junípero Serra in an area long inhabited by the Kumeyaay people. The mission and the surrounding area were named for the Catholic saint Didacus of Alcalá, a Spaniard more commonly known as San Diego. The mission was the site of the first Christian burial in Alta California. The original mission burned in 1775 during an uprising by local natives San Diego is also generally regarded as the site of the region's first public execution, in 1778. Father Luis Jayme, California's first Christian martyr who was among those killed during the 1775 uprising against the mission, lies entombed beneath the chancel floor. The current church, built in the early 19th century, is the fifth to stand on this location. The mission site is a National Historic Landmark.
History
The former Spanish settlement at the Kumeyaay Nipawai lies within that area occupied during the late Paleoindian period and continuing on into the present day by the Native society commonly known as the Diegueño; the name denotes those people who were ministered by the padres at Mission San Diego de Alcalá. Relatively, much is known about the native inhabitants in recent centuries, thanks in part to the efforts of Juan Rodríguez Cabrillo, who explored the coast in the service of Spain. He documented his in the coastal villages he encountered along the Southern California coast in October 1542. Cabrillo was the first European to set foot in what is now the state of California and the first to encounter San Diego Bay. On the evening of September 28, 1542, the ships San Salvador and Victoria sailed into the harbor, whereupon Cabrillo christened it "San Miguel." During that expedition a landing party went ashore and briefly interacted with a small group of natives. Some sixty years later another Spanish explorer, Sebastián Vizcaíno, made landfall some ten miles from the present Mission site. Under Vizcaíno's command the San Diego, Santo Tomás, and frigate Tres Reyes dropped anchor on November 10, 1602, and the port was renamed "San Diego de Alcalá." It would be another 167 years before the Spanish returned to San Diego.
Ever since the voyages of Christopher Columbus, the Kingdom of Spain sought to establish missions to convert the pagans in Nueva España ("New Spain") to Roman Catholicism to save souls and to facilitate colonization of these lands. However, it was not until 1741—the time of the Vitus Bering expedition, when the territorial ambitions of Tsarist Russia towards North America became known—that King Philip V felt such installations were necessary in Upper California. In 1769, Visitador General José de Gálvez sent the expedition of Junípero Serra and Gaspar de Portolà to found missions and presidios at San Diego and Monterey (see Timeline of the Portolà expedition), thereby securing Spain's claim to the Pacific Coast harbors recommended by Cabrillo and Vizcaino. Two groups traveled from Lower California on foot, while a pair of packet ships (bearing supplies) traveled up the coast from the Baja California Peninsula.
The original mission church was destroyed by fire in 1775 during an uprising by local indigenous people. An estimated 800 "American Indians" pillaged the mission, burned it to the ground and massacred a blacksmith, a carpenter (mortally wounded), and Father Jayme, who became California's first Catholic Martyr. Father Jayme is buried next to the altar in the present church. Survivors of the night-long attack were one corporal and three Leather Jacket soldiers, one blacksmith, two children who were the son and nephew of the Presidio commandant, and Associate Pastor Father Vicente Fuster. Father Serra then returned to the site of the original structure in 1776 to oversee the rebuilding of the mission. However, the scarce amount of water and the difficulty of making the land till (feasible for preparing crops to be planted by plowing and fertilizing) made the re-establishment of the mission a long and difficult process. From 1778 to 1795, the mission focused on horse and mule breeding, providing other missions in Alta and Baja California an average of 16 animals per year. After it was restored as a Catholic mission, Mission San Diego de Alcalá saw a record number of spiritual results recorded when 567 baptisms were performed, and neophytes numbered 908.
Rancho period (1834–1849)
On August 9, 1834, Governor Figueroa issued his "Decree of Confiscation." The missions were offered for sale to citizens, who were unable to come up with the price, so all mission property was broken up into ranchos and given to ex-military officers who had fought in the War of Independence against Spain. On June 8, 1846, Mission San Diego de Alcalá was given to Santiago Argüello by Governor Pío Pico "...for services rendered to the government." After the United States annexed California, the Mission was used by the military from 1846 to 1862.
California statehood (1850–1900)
President Abraham Lincoln signed a proclamation on May 23, 1862, that restored ownership of the Mission proper to the Roman Catholic Church. When Mission San Diego de Alcalá was granted back to the Church, it was in ruins. In the 1880s Father Anthony Ubach began to restore the old Mission buildings.
20th century and beyond (1901 – present)
Father Ubach died in 1907 and restoration work ceased until 1931. In 1941, the Mission once again became a parish church, in what is still an active parish serving the Diocese of San Diego. In 1976, Pope Paul VI designated the Mission church as a minor basilica.
Mission San Diego de Alcalá is located within San Diego city limits, near the intersection of Interstate 8 and Interstate 15, and approximately one mile east of SDCCU Stadium.
Other historic designations
California Historical Landmark #242
California Historical Landmark #784 – El Camino Real (starting point in Alta California)
City of San Diego Historic Designation #113
Mission industries
The goal of the missions was, above all, to become self-sufficient in relatively short order. Farming, therefore, was the most important industry of any mission. Prior to the establishment of the missions, the native peoples knew how to utilize bone, seashells, stone, and wood for building, tool making, weapons, and much more. The missionaries discovered that the Indians, who regarded labor as degrading to men, had to be taught industry in order to learn how to be self-supportive. The result was the establishment of a manual training school that comprised agriculture, the mechanical arts, and the raising and care of livestock. Everything consumed and otherwise utilized by the natives was produced at the missions under the supervision of the padres; thus, the neophytes not only supported themselves, but after 1811 sustained the entire military and civil government of California. Wheat, corn, wine grapes, barley, beans, cattle, horses, and sheep were the major crops at San Diego. In 1795, construction on a system of aqueducts was begun to bring water to the fields and the Mission (the first irrigation project in Upper California). The building manager was Fray Pedro Panto, who was poisoned by his Indian cook Nazario before the project was completed. In his testimony, in the trial that followed, Nazario stated that he had poisoned the friar due to constant beatings inflicted by Friar Panto.
The Mission San Diego was primarily supported from lands included in a Spanish royal land grant, encompassing roughly the eastern third of the current City of San Diego, as well as most of the cities of La Mesa and Lemon Grove. While not exact, its boundaries are roughly Interstate 805, Miramar Road, Route 125 (California), Skyline Drive and Division Street. Boundary Street (which parallels Interstate 805 in the neighborhood of North Park) lies directly on one portion of the boundary and draws its name from it.
Wine
The first vineyards planted in California were planted at the mission in 1769. Father Junípero Serra planted these initial vines. The vines were of the Mission variety, which were brought to Mexico in the 16th Century. Unfortunately, the initial plantings did not survive, with Mission San Juan Capistrano having the first surviving plantings. By at least 1781, wine was being produced at the mission. One source claims that the vineyard on mission lands reached up to fifty thousand acres. The mission vineyards were around until at least 1823.
Mission bells
Bells were vitally important to daily life at any mission. The bells were rung at mealtimes, to call the Mission residents to work and to religious services, during births and funerals, to signal the approach of a ship or returning missionary, and at other times; novices were instructed in the intricate rituals associated with ringing the mission bells. This mission had five bells. The first bell was hung up on a tree
Gallery
See also
Spanish missions in California
List of Spanish missions in California
Santa Ysabel Asistencia
Presidio of San Diego
USNS Mission San Diego (AO-121) – a Mission Buenaventura Class fleet oiler built during World War II.
Union Station (San Diego)
List of the oldest buildings in the United States
Notes
References
External links
Official Mission San Diego de Alcala website
"Sociopolitical Aspects of the 1775 Revolt at Mission San Diego de Alcalá: an Ethnohistorical Approach" by Richard L. Carrico
Spanish Wiki page with Fray Pedro Panto's biography
Early photographs, sketches of Mission San Diego de Alcala, via Calisphere, California Digital Library
Early History of the California Coast, a National Park Service Discover Our Shared Heritage Travel Itinerary
Listing, photographs, and drawings at the Historic American Buildings Survey
j
San Diego de Alcala
1769 in Alta California
Museums in San Diego
Religious museums in California
Diego de Alcala
California Historical Landmarks
History of San Diego
History of San Diego County, California
Mission Valley, San Diego
San Diego River
National Historic Landmarks in California
National Register of Historic Places in San Diego
Churches on the National Register of Historic Places in California
1769 establishments in Alta California
Native American history of California
San Antonio–San Diego Mail Line
Stagecoach stops in the United States
Churches in San Diego
Junípero Serra
Spanish Colonial architecture in the United States |
null | null | Ionization | eng_Latn | Ionization or ionisation is the process by which an atom or a molecule acquires a negative or positive charge by gaining or losing electrons, often in conjunction with other chemical changes. The resulting electrically charged atom or molecule is called an ion. Ionization can result from the loss of an electron after collisions with subatomic particles, collisions with other atoms, molecules and ions, or through the interaction with electromagnetic radiation. Heterolytic bond cleavage and heterolytic substitution reactions can result in the formation of ion pairs. Ionization can occur through radioactive decay by the internal conversion process, in which an excited nucleus transfers its energy to one of the inner-shell electrons causing it to be ejected.
Uses
Everyday examples of gas ionization are such as within a fluorescent lamp or other electrical discharge lamps. It is also used in radiation detectors such as the Geiger-Müller counter or the ionization chamber. The ionization process is widely used in a variety of equipment in fundamental science (e.g., mass spectrometry) and in industry (e.g., radiation therapy).
Production of ions
Negatively charged ions are produced when a free electron collides with an atom and is subsequently trapped inside the electric potential barrier, releasing any excess energy. The process is known as electron capture ionization.
Positively charged ions are produced by transferring an amount of energy to a bound electron in a collision with charged particles (e.g. ions, electrons or positrons) or with photons. The threshold amount of the required energy is known as ionization potential. The study of such collisions is of fundamental importance with regard to the few-body problem, which is one of the major unsolved problems in physics. Kinematically complete experiments, i.e. experiments in which the complete momentum vector of all collision fragments (the scattered projectile, the recoiling target-ion, and the ejected electron) are determined, have contributed to major advances in the theoretical understanding of the few-body problem in recent years.
Adiabatic ionization
Adiabatic ionization is a form of ionization in which an electron is removed from or added to an atom or molecule in its lowest energy state to form an ion in its lowest energy state.
The Townsend discharge is a good example of the creation of positive ions and free electrons due to ion impact. It is a cascade reaction involving electrons in a region with a sufficiently high electric field in a gaseous medium that can be ionized, such as air. Following an original ionization event, due to such as ionizing radiation, the positive ion drifts towards the cathode, while the free electron drifts towards the anode of the device. If the electric field is strong enough, the free electron gains sufficient energy to liberate a further electron when it next collides with another molecule. The two free electrons then travel towards the anode and gain sufficient energy from the electric field to cause impact ionization when the next collisions occur; and so on. This is effectively a chain reaction of electron generation, and is dependent on the free electrons gaining sufficient energy between collisions to sustain the avalanche.
Ionization efficiency is the ratio of the number of ions formed to the number of electrons or photons used.
Ionization energy of atoms
The trend in the ionization energy of atoms is often used to demonstrate the periodic behavior of atoms with respect to the atomic number, as summarized by ordering atoms in Mendeleev's table. This is a valuable tool for establishing and understanding the ordering of electrons in atomic orbitals without going into the details of wave functions or the ionization process. An example is presented in the figure at right. The periodic abrupt decrease in ionization potential after rare gas atoms, for instance, indicates the emergence of a new shell in alkali metals. In addition, the local maximums in the ionization energy plot, moving from left to right in a row, are indicative of s, p, d, and f sub-shells.
Semi-classical description of ionization
Classical physics and the Bohr model of the atom can qualitatively explain photoionization and collision-mediated ionization. In these cases, during the ionization process, the energy of the electron exceeds the energy difference of the potential barrier it is trying to pass. The semi-classical description, however, cannot describe tunnel ionization since the process involves the passage of electron through a classically forbidden potential barrier.
Quantum mechanical description of ionization
The interaction of atoms and molecules with sufficiently strong laser pulses leads to the ionization to singly or multiply charged ions. The ionization rate, i.e. the ionization probability in unit time, can only be calculated using quantum mechanics. In general, the analytic solutions are not available, and the approximations required for manageable numerical calculations do not provide accurate enough results. However, when the laser intensity is sufficiently high, the detailed structure of the atom or molecule can be ignored and analytic solution for the ionization rate is possible.
Tunnel ionization
Tunnel ionization is ionization due to quantum tunneling. In classical ionization, an electron must have enough energy to make it over the potential barrier, but quantum tunneling allows the electron simply to go through the potential barrier instead of going all the way over it because of the wave nature of the electron. The probability of an electron's tunneling through the barrier drops off exponentially with the width of the potential barrier. Therefore, an electron with a higher energy can make it further up the potential barrier, leaving a much thinner barrier to tunnel through and, thus, a greater chance to do so. In practice, tunnel ionization is observable when the atom or molecule is interacting with near-infrared strong laser pulses. This process can be understood as a process by which a bounded electron, through the absorption of more than one photon from the laser field, is ionized. This picture is generally known as multiphoton ionization (MPI).
Keldysh modeled the MPI process as a transition of the electron from the ground state of the atom to the Volkov states. In this model the perturbation of the ground state by the laser field is neglected and the details of atomic structure in determining the ionization probability are not taken into account. The major difficulty with Keldysh's model was its neglect of the effects of Coulomb interaction on the final state of the electron. As it is observed from figure, the Coulomb field is not very small in magnitude compared to the potential of the laser at larger distances from the nucleus. This is in contrast to the approximation made by neglecting the potential of the laser at regions near the nucleus. Perelomov et al. included the Coulomb interaction at larger internuclear distances. Their model (which we call PPT model) was derived for short range potential and includes the effect of the long range Coulomb interaction through the first order correction in the quasi-classical action. Larochelle et al. have compared the theoretically predicted ion versus intensity curves of rare gas atoms interacting with a Ti:Sapphire laser with experimental measurement. They have shown that the total ionization rate predicted by the PPT model fit very well the experimental ion yields for all rare gases in the intermediate regime of Keldysh parameter.
The rate of MPI on atom with an ionization potential in a linearly polarized laser with frequency is given by
where
is the Keldysh's adiabaticity parameter,
,
is the peak electric field of laser and
.
The coefficients , and are given by
The coefficient is given by
where
Quasi-static tunnel ionization
The quasi-static tunnelling (QST) is the ionization whose rate can be satisfactorily predicted by the ADK model, i.e. the limit of the PPT model when approaches zero. The rate of QST is given by
As compared to the absence of summation over n, which represent different above threshold ionization (ATI) peaks, is remarkable.
Strong field approximation for the ionization rate
The calculations of PPT are done in the E-gauge, meaning that the laser field is taken as electromagnetic waves. The ionization rate can also be calculated in A-gauge, which emphasizes the particle nature of light (absorbing multiple photons during ionization). This approach was adopted by Krainov model based on the earlier works of Faisal and Reiss. The resulting rate is given by
where:
with being the ponderomotive energy,
is the minimum number of photons necessary to ionize the atom,
is the double Bessel function,
with the angle between the momentum of the electron, p, and the electric field of the laser, F,
FT is the three-dimensional Fourier transform, and
incorporates the Coulomb correction in the SFA model.
Atomic stabilization/population trapping
In calculating the rate of MPI of atoms only transitions to the continuum states are considered. Such an approximation is acceptable as long as there is no multiphoton resonance between the ground state and some excited states. However, in real situation of interaction with pulsed lasers, during the evolution of laser intensity, due to different Stark shift of the ground and excited states there is a possibility that some excited state go into multiphoton resonance with the ground state. Within the dressed atom picture, the ground state dressed by photons and the resonant state undergo an avoided crossing at the resonance intensity . The minimum distance, , at the avoided crossing is proportional to the generalized Rabi frequency, coupling the two states. According to Story et al., the probability of remaining in the ground state, , is given by
where is the time-dependent energy difference between the two dressed states. In interaction with a short pulse, if the dynamic resonance is reached in the rising or the falling part of the pulse, the population practically remains in the ground state and the effect of multiphoton resonances may be neglected. However, if the states go onto resonance at the peak of the pulse, where , then the excited state is populated. After being populated, since the ionization potential of the excited state is small, it is expected that the electron will be instantly ionized.
In 1992, de Boer and Muller showed that Xe atoms subjected to short laser pulses could survive in the highly excited states 4f, 5f, and 6f. These states were believed to have been excited by the dynamic Stark shift of the levels into multiphoton resonance with the field during the rising part of the laser pulse. Subsequent evolution of the laser pulse did not ionize completely these states leaving behind some highly excited atoms. We shall refer to this phenomenon as "population trapping".
We mention the theoretical calculation that incomplete ionization occurs whenever there is parallel resonant excitation into a common level with ionization loss. We consider a state such as 6f of Xe which consists of 7 quasi-degnerate levels in the range of the laser bandwidth. These levels along with the continuum constitute a lambda system. The mechanism of the lambda type trapping is schematically presented in figure. At the rising part of the pulse (a) the excited state (with two degenerate levels 1 and 2) are not in multiphoton resonance with the ground state. The electron is ionized through multiphoton coupling with the continuum. As the intensity of the pulse is increased the excited state and the continuum are shifted in energy due to the Stark shift. At the peak of the pulse (b) the excited states go into multiphoton resonance with the ground state. As the intensity starts to decrease (c), the two state are coupled through continuum and the population is trapped in a coherent superposition of the two states. Under subsequent action of the same pulse, due to interference in the transition amplitudes of the lambda system, the field cannot ionize the population completely and a fraction of the population will be trapped in a coherent superposition of the quasi degenerate levels. According to this explanation the states with higher angular momentum- with more sublevels- would have a higher probability of trapping the population. In general the strength of the trapping will be determined by the strength of the two photon coupling between the quasi-degenerate levels via the continuum. In 1996, using the very stable laser and by minimizing the masking effects of the focal region expansion with increasing intensity, Talebpour et al. observed structures on the curves of singly charged ions of Xe, Kr and Ar. These structures were attributed to electron trapping in the strong laser field. A more unambiguous demonstration of population trapping has been reported by T. Morishita and C. D. Lin.
Non-sequential multiple ionization
The phenomenon of non-sequential ionization (NSI) of atoms exposed to intense laser fields has been a subject of many theoretical and experimental studies since 1983. The pioneering work began with the observation of a “knee” structure on the Xe2+ ion signal versus intensity curve by L’Huillier et al. From the experimental point of view, the NS double ionization refers to processes which somehow enhance the rate of production of doubly charged ions by a huge factor at intensities below the saturation intensity of the singly charged ion. Many, on the other hand, prefer to define the NSI as a process by which two electrons are ionized nearly simultaneously. This definition implies that apart from the sequential channel there is another channel which is the main contribution to the production of doubly charged ions at lower intensities. The first observation of triple NSI in argon interacting with a 1 µm laser was reported by Augst et al. Later, systematically studying the NSI of all rare gas atoms, the quadruple NSI of Xe was observed. The most important conclusion of this study was the observation of the following relation between the rate of NSI to any charge state and the rate of tunnel ionization (predicted by the ADK formula) to the previous charge states;
where is the rate of quasi-static tunneling to i'th charge state and are some constants depending on the wavelength of the laser (but not on the pulse duration).
Two models have been proposed to explain the non-sequential ionization; the shake-off model and electron re-scattering model. The shake-off (SO) model, first proposed by Fittinghoff et al., is adopted from the field of ionization of atoms by X rays and electron projectiles where the SO process is one of the major mechanisms responsible for the multiple ionization of atoms. The SO model describes the NS process as a mechanism where one electron is ionized by the laser field and the departure of this electron is so rapid that the remaining electrons do not have enough time to adjust themselves to the new energy states. Therefore, there is a certain probability that, after the ionization of the first electron, a second electron is excited to states with higher energy (shake-up) or even ionized (shake-off). We should mention that, until now, there has been no quantitative calculation based on the SO model, and the model is still qualitative.
The electron rescattering model was independently developed by Kuchiev, Schafer et al, Corkum, Becker and Faisal and Faisal and Becker. The principal features of the model can be understood easily from Corkum's version. Corkum's model describes the NS ionization as a process whereby an electron is tunnel ionized. The electron then interacts with the laser field where it is accelerated away from the nuclear core. If the electron has been ionized at an appropriate phase of the field, it will pass by the position of the remaining ion half a cycle later, where it can free an additional electron by electron impact. Only half of the time the electron is released with the appropriate phase and the other half it never return to the nuclear core. The maximum kinetic energy that the returning electron can have is 3.17 times the ponderomotive potential () of the laser. Corkum's model places a cut-off limit on the minimum intensity ( is proportional to intensity) where ionization due to re-scattering can occur.
The re-scattering model in Kuchiev's version (Kuchiev's model) is quantum mechanical. The basic idea of the model is illustrated by Feynman diagrams in figure a. First both electrons are in the ground state of an atom. The lines marked a and b describe the corresponding atomic states. Then the electron a is ionized. The beginning of the ionization process is shown by the intersection with a sloped dashed line. where the MPI occurs. The propagation of the ionized electron in the laser field, during which it absorbs other photons (ATI), is shown by the full thick line. The collision of this electron with the parent atomic ion is shown by a vertical dotted line representing the Coulomb interaction between the electrons. The state marked with c describes the ion excitation to a discrete or continuum state. Figure b describes the exchange process. Kuchiev's model, contrary to Corkum's model, does not predict any threshold intensity for the occurrence of NS ionization.
Kuciev did not include the Coulomb effects on the dynamics of the ionized electron. This resulted in the underestimation of the double ionization rate by a huge factor. Obviously, in the approach of Becker and Faisal (which is equivalent to Kuchiev's model in spirit), this drawback does not exist. In fact, their model is more exact and does not suffer from the large number of approximations made by Kuchiev. Their calculation results perfectly fit with the experimental results of Walker et al. Becker and Faisal have been able to fit the experimental results on the multiple NSI of rare gas atoms using their model. As a result, the electron re-scattering can be taken as the main mechanism for the occurrence of the NSI process.
Multiphoton ionization of inner-valence electrons and fragmentation of polyatomic molecules
The ionization of inner valence electrons are responsible for the fragmentation of polyatomic molecules in strong laser fields. According to a qualitative model the dissociation of the molecules occurs through a three-step mechanism:
MPI of electrons from the inner orbitals of the molecule which results in a molecular ion in ro-vibrational levels of an excited electronic state;
Rapid radiationless transition to the high-lying ro-vibrational levels of a lower electronic state; and
Subsequent dissociation of the ion to different fragments through various fragmentation channels.
The short pulse induced molecular fragmentation may be used as an ion source for high performance mass spectroscopy. The selectivity provided by a short pulse based source is superior to that expected when using the conventional electron ionization based sources, in particular when the identification of optical isomers is required.
Kramers-Henneberger frame and ionization phase effects
Studying the strong field ionization of the atom in so called Kramers-Henneberger (K-H) frame leads to the conclusion that the ionization efficiency strongly depends on the temporal details of the ionizing pulse but not necessarily on the field strength and the total energy of the ionizing pulse pumped into the atom. The Kramers-Henneberger frame is the non-intertial frame moving with the free electron under the influence of the harmonic laser pulse. The free electron solution of the Newton equations for the electron in one dimension in the harmonic laser field
will be also harmonic
The frame comoving with this electron will be obtained by the coordinate transformation
while the added Coulomb potential will be
The full cycle time-average of that potential which is
will be the even function of and therefore having the maximum at while for that initial condition the solution will be in the K-H and it will be therefore identical to the free electron solution in the laboratory frame. The electron velocity on the other hand is phase shifted both to the field strength and to the electron position:
Therefore, considering the wavelet pulses and defining the ionization as the full escape from the line segment of the length 2r (or from the spherical region in three dimensions) the full ionization happens in the classical model after the time or no ionization at all depending if the harmonic field wavelet is cut at the zero minimum or the maximum velocity.
Dissociation – distinction
A substance may dissociate without necessarily producing ions. As an example, the molecules of table sugar dissociate in water (sugar is dissolved) but exist as intact neutral entities. Another subtle event is the dissociation of sodium chloride (table salt) into sodium and chlorine ions. Although it may seem as a case of ionization, in reality the ions already exist within the crystal lattice. When salt is dissociated, its constituent ions are simply surrounded by water molecules and their effects are visible (e.g. the solution becomes electrolytic). However, no transfer or displacement of electrons occurs. Actually, the chemical synthesis of salt involves ionization. This is a chemical reaction.
See also
Above threshold ionization
Ionization chamber – Instrument for detecting gaseous ionization, used in ionizing radiation measurements
Ion source
Photoionization
Thermal ionization
Electron ionization
Chemical ionization
Townsend avalanche – The chain reaction of ionization occurring in a gas with an applied electric field
Table
References
External links
Ions
Molecular physics
Atomic physics
Physical chemistry
Quantum chemistry
Mass spectrometry |
null | null | Group of Eight | eng_Latn | The Group of Eight (G8) was an inter-governmental political forum from 1997 until 2014. It had formed from incorporating the country of Russia into the Group of Seven, or G7, and returned to its previous name after Russia was removed in 2014.
The forum originated with a 1975 summit hosted by France that brought together representatives of six governments: France, Germany, Italy, Japan, the United Kingdom, and the United States, thus leading to the name Group of Six or G6. The summit came to be known as the Group of Seven in 1976 with the addition of Canada. Russia was added to the political forum from 1997, which the following year became known as the G8. In March 2014 Russia was suspended indefinitely following the annexation of Crimea, whereupon the political forum name reverted to G7. In 2017, Russia announced its permanent withdrawal from the G8. However, several representatives of G7 countries stated that they would be interested in Russia's return to the group. The European Union was represented at the G8 since the 1980s as a "nonenumerated" participant, but originally could not host or chair summits. The 40th summit was the first time the European Union was able to host and chair a summit. Collectively, in 2012 the G8 nations comprised 50.1 percent of 2012 global nominal GDP and 40.9 percent of global GDP (PPP).
"G7" can refer to the member states in aggregate or to the annual summit meeting of the G7 heads of government. G7 ministers also meet throughout the year, such as the G7 finance ministers (who meet four times a year), G7 foreign ministers, or G7 environment ministers.
Each calendar year, the responsibility of hosting the G8 was rotated through the member states in the following order: France, United States, United Kingdom, Russia (suspended), Germany, Japan, Italy, and Canada. The holder of the presidency sets the agenda, hosts the summit for that year, and determines which ministerial meetings will take place.
In 2005, the UK government initiated the practice of inviting five leading emerging markets – Brazil, China, India, Mexico, and South Africa – to participate in the G8 meetings that came to be known as G8+5, but this practice was short-lived. With the G20 major economies growing in stature since the 2008 Washington summit, world leaders from the group announced at their Pittsburgh summit in September 2009 that the group would replace the G8 as the main economic council of wealthy nations. Nevertheless, the G7 retains its relevance as a "steering group for the West", with special significance appointed to Japan.
History
Following 1994's G7 summit in Naples, Russian officials held separate meetings with leaders of the G7 after the group's summits. This informal arrangement was dubbed the Political 8 (P8)—or, colloquially, the G7+1. At the invitation of UK Prime Minister Tony Blair and U.S. President Bill Clinton, President Boris Yeltsin was invited first as a guest observer, later as a full participant. It was seen as a way to encourage Yeltsin with his capitalist reforms. Russia formally joined the group in 1998, resulting in the Group of Eight, or G8.
Food
A major focus of the G8 since 2009 has been the global supply of food. At the 2009 L'Aquila summit, the G8's members promised to contribute $22 billion to the issue. By 2015, 93% of funds had been disbursed to projects like sustainable agriculture development and adequate emergency food aid assistance.
At the 2012 summit, President Barack Obama asked G8 leaders to adopt the New Alliance for Food Security and Nutrition initiative to "help the rural poor produce more food and sell it in thriving local and regional markets as well as on the global market". Ghana became one of the first six African countries to sign up to the G8 New Alliance for Food Security and Nutrition in 2012. There was, however, almost no knowledge of the G8 initiative among some stakeholders, including farmers, academics and agricultural campaign groups. Confusion surrounding the plans was made worse, critics say, by "a dizzying array of regional and national agriculture programmes that are inaccessible to ordinary people."
Russia's participation suspension (2014)
On 24 March 2014, the G7 members cancelled the planned G8 summit that was to be held in June of that year in the Russian city of Sochi, and suspended Russia's membership of the group, due to Russia's annexation of Crimea; nevertheless, they stopped short of outright permanent expulsion. Russian foreign minister Sergei Lavrov downplayed the importance of the decision by the U.S. and its allies, and pointed out that major international decisions were made by the G20 countries.
Later on, the Italian Foreign Affairs minister Federica Mogherini and other Italian authorities, along with the EastWest Institute board member Wolfgang Ischinger, suggested that Russia may restore its membership in the group. In April 2015, the German foreign minister Frank-Walter Steinmeier said that Russia would be welcomed to return to G8 provided the Minsk Protocol was implemented. In 2016, he added that "none of the major international conflicts can be solved without Russia", and the G7 countries will consider Russia's return to the group in 2017. The same year, Japanese Prime Minister Shinzō Abe called for Russia's return to G8, stating that Russia's involvement is "crucial to tackling multiple crises in the Middle East". In January 2017, the Italian foreign minister Angelino Alfano said that Italy hopes for "resuming the G8 format with Russia and ending the atmosphere of the Cold War". On 13 January 2017, Russia announced that it would permanently leave the G8 grouping. Nonetheless, Christian Lindner, the leader of Free Democratic Party of Germany and member of the Bundestag, said that Putin should be "asked to join the table of the G7" so that one could "talk with him and not about him", and "we cannot make all things dependent on the situation in Crimea". In April 2018, the German politicians and members of the Bundestag Sahra Wagenknecht and Alexander Graf Lambsdorff said that Russia should be invited back to the group and attend the 2018 summit in Canada: "Russia should again be at the table during the [June] summit at the latest" because "peace in Europe and also in the Middle East is only possible with Russia". The US President Donald Trump also stated that Russia should be reinstated to the group; his appeal was supported by the Italian Prime Minister Giuseppe Conte. After several G7 members quickly rejected US President Trump suggestion to accept again Russian Federation in the G8, Russian Foreign Minister Lavrov said the Russian Federation wasn't interested in rejoining the political forum. He also said that G20 is sufficient for Russian Federation. In the final statement of the 2018 meeting in Canada, the G7 members announced to continue sanctions and also to be ready to take further restrictive measures against the Russian Federation for the failure of Minsk Agreement complete implementation.
Structure and activities
By design, the G8 deliberately lacked an administrative structure like those for international organizations, such as the United Nations or the World Bank. The group does not have a permanent secretariat, or offices for its members.
The presidency of the group rotates annually among member countries, with each new term beginning on 1 January of the year. The rotation order is: France, the United States, the United Kingdom, Russia (suspended), Germany, Japan, Italy, and Canada. The country holding the presidency is responsible for planning and hosting a series of ministerial-level meetings, leading up to a mid-year summit attended by the heads of government. The president of the European Commission participates as an equal in all summit events.
The ministerial meetings bring together ministers responsible for various portfolios to discuss issues of mutual or global concern. The range of topics include health, law enforcement, labor, economic and social development, energy, environment, foreign affairs, justice and interior, terrorism, and trade. There are also a separate set of meetings known as the G8+5, created during the 2005 Gleneagles, Scotland summit, that is attended by finance and energy ministers from all eight member countries in addition to the five "outreach countries" which are also known as the Group of Five—Brazil, People's Republic of China, India, Mexico, and South Africa.
In June 2005, justice ministers and interior ministers from the G8 countries agreed to launch an international database on pedophiles. The G8 officials also agreed to pool data on terrorism, subject to restrictions by privacy and security laws in individual countries.
Global energy
At the Heiligendamm Summit in 2007, the G8 acknowledged a proposal from the EU for a worldwide initiative on efficient energy use. They agreed to explore, along with the International Energy Agency, the most effective means to promote energy efficiency internationally. A year later, on 8 June 2008, the G8 along with China, India, South Korea and the European Community established the International Partnership for Energy Efficiency Cooperation, at the Energy Ministerial meeting hosted by Japan holding 2008 G8 Presidency, in Aomori.
G8 Finance Ministers, whilst in preparation for the 34th Summit of the G8 Heads of State and Government in Toyako, Hokkaido, met on the 13 and 14 June 2008, in Osaka, Japan. They agreed to the "G8 Action Plan for Climate Change to Enhance the Engagement of Private and Public Financial Institutions." In closing, Ministers supported the launch of new Climate Investment Funds (CIFs) by the World Bank, which will help existing efforts until a new framework under the UNFCCC is implemented after 2012. The UNFCCC is not on track to meeting any of its stated goals.
In July 2005, the G8 Summit endorsed the IPHE in its Plan of Action on Climate Change, Clean Energy and Sustainable Development, and identified it as a medium of cooperation and collaboration to develop clean energy technologies.
Annual summit
The first G8 summit was held in 1997 after Russia formally joined the G7 group, and the last one was held in 2013. The 2014 summit was scheduled to be held in Russia. However, due to the Crimean crisis, the other seven countries decided to hold a separate meeting without Russia as a G7 summit in Brussels, Belgium.
Criticism
One type of criticism is that members of G8 do not do enough to help global problems, due to strict patent policy and other issues related to globalization. In Unraveling Global Apartheid, political analyst Titus Alexander described the G7, as it was in 1996, as the 'cabinet' of global minority rule, with a coordinating role in world affairs.
In 2012 The Heritage Foundation criticized the G8 for advocating food security without making room for economic freedom.
Relevance
The G8's relevance has been subject to debate from 2008 onward. It represented the major industrialized countries but critics argued that the G8 no longer represented the world's most powerful economies, as China has surpassed every economy but the United States.
Vladimir Putin did not attend the 2012 G8 summit at Camp David, causing Foreign Policy magazine to remark that the summit has generally outlived its usefulness as a viable international gathering of foreign leaders. Two years later, Russia was suspended from the G8, then chose to leave permanently in January 2017.
The G20 major economies leaders' summit has had an increased level of international prestige and influence. However, British Prime Minister David Cameron said of the G8 in 2012:
Some people ask, does the G8 still matter, when we have a Group of 20? My answer is, yes. The G8 is a group of like-minded countries that share a belief in free enterprise as the best route to growth. As eight countries making up about half the world's gross domestic product, the standards we set, the commitments we make, and the steps we take can help solve vital global issues, fire up economies and drive prosperity all over the world.
Youth 8 Summit
The Y8 Summit or simply Y8, formerly known as the G8 Youth Summit is the youth counterpart to the G8 summit. The summits were organized from 2006 to 2013. The first summit to use the name Y8 took place in May 2012 in Puebla, Mexico, alongside the Youth G8 that took place in Washington, D.C. the same year. From 2016 onwards, similar youth conferences were organized under the name Y7 Summit.
The Y8 Summit brings together young leaders from G8 nations and the European Union to facilitate discussions of international affairs, promote cross-cultural understanding, and build global friendships. The conference closely follows the formal negotiation procedures of the G8 Summit. The Y8 Summit represents the innovative voice of young adults between the age of 18 and 35. At the end of the summit, the delegates jointly come up with a consensus-based written statement, the Final Communiqué. This document is subsequently presented to G8 leaders in order to inspire positive change.
The Y8 Summit was organized annually by a global network of youth-led organizations called The IDEA (The International Diplomatic Engagement Association). The organizations undertake the selection processes for their respective national delegations, while the hosting country is responsible for organizing the summit. An example of such a youth-led organization is the Young European Leadership association, which recruits and sends EU Delegates.
The goal of the Y8 Summit is to bring together young people from around the world to allow the voices and opinions of young generations to be heard and to encourage them to take part in global decision-making processes.
* The Y8 Summit 2014 in Moscow was suspended due to the suspension of Russia from the G8.
See also
D-8 Organization for Economic Cooperation
Forum for the Future (Bahrain 2005)
G3 Free Trade Agreement
G4 (EU)
G-20 major economies
Group of Two
Group of Seven
Group of Eleven
Group of 15
Group of 24
Group of 30
Junior 8
List of countries by GDP (nominal)
List of countries by military expenditures
List of country groupings
List of G8 leaders
List of G8 summit resorts
List of longest serving G8 leaders
List of multilateral free-trade agreements
North–South divide
Western Bloc
Great power
World Social Forum
References
Further reading
Bayne, Nicholas and Robert D. Putnam. (2000). Hanging in There: The G7 and G8 Summit in Maturity and Renewal. Aldershot, Hampshire: Ashgate Publishing. ;
Haas, P.M. (1992). "Introduction. Epistemic communities and international policy coordination," International Organization 46, 1:1–35.
Hajnal, Peter I. (1999). The G8 system and the G20 : Evolution, Role and Documentation. Aldershot, Hampshire: Ashgate Publishing. ;
Kokotsis, Eleonore. (1999). Keeping International Commitments: Compliance, Credibility, and the G7, 1988–1995. New York: Garland Publishing. ;
Reinalda, Bob and Bertjan Verbeek. (1998). Autonomous Policy Making by International Organizations. London: Routledge. ;
External links
G8 Information Centre, G8 Research Group, University of Toronto
"Special Report: G8", Guardian Unlimited
"Profile: G8", BBC News
"We are deeply concerned. Again", New Statesman, 4 July 2005, —G8 development concerns since 1977
G8 Information Centre Finance Ministers Meetings
"G8: Cooking the books won’t feed anyone", Oxfam International
"Dear G8 Leaders, don’t lie about your aid", Oxfam International Blogs
"Wait, the G-8 still exists?", Foreign Policy Magazine
"Is this the last G-8 summit meeting?", Foreign Policy Magazine
Chronique ONU | Le Conseil économique et social, Le Groupe des huit et le PARADOXE CONSTITUTIONNEL "The Group of Eight, ECOSOC and the Constitutional Paradox"
No. of G8 Summit Protestors (1998–2015) Katapult-Magazin
Economic country classifications
Defunct intergovernmental organizations
International economic organizations
20th-century diplomatic conferences
21st-century diplomatic conferences
Organizations established in 1997
Organizations disestablished in 2014 |
null | null | Mission San Juan Bautista | eng_Latn | Mission San Juan Bautista is a Spanish mission in San Juan Bautista, San Benito County, California. Founded on June 24, 1797 by Fermín Lasuén of the Franciscan order, the mission was the fifteenth of the Spanish missions established in present-day California. Named for Saint John the Baptist, the mission is the namesake of the city of San Juan Bautista.
Barracks for the soldiers, a nunnery, the Jose Castro House, and other buildings were constructed around a large grassy plaza in front of the church and can be seen today in their original form. The Ohlone, the original residents of the valley, were brought to live at the mission and baptized, followed by Yokuts from the Central Valley. Mission San Juan Bautista has served mass daily since 1797, and today functions as a parish church of the Diocese of Monterey.
History
Following its creation in 1797, San Juan's population grew quickly. By 1803, there were 1,036 Native Americans living at the mission. Ranching and farming activity had moved apace, with 1,036 cattle, 4,600 sheep, 22 swine, 540 horses and 8 mules counted that year. At the same time, the harvest of wheat, barley and corn was estimated at 2,018 fanegas, each of about 220 pounds.
Father Pedro Estévan Tápis (who had a special talent for music) joined Father Felipe Arroyo de la Cuesta, at Mission San Juan Bautista in 1815 to teach singing to the Native Americans. He employed a system of notation developed in Spain that uses varied colors or textures for polyphonic music, usually (from bottom to top) solid black, solid red, black outline (sometimes solid yellow) and red outline (or black outline when yellow was used). His choir of Native American boys performed for many visitors, earning the San Juan Bautista Mission the nickname "the Mission of Music." Two of his handwritten choir books are preserved at the San Juan Bautista Museum. When Father Tapis died in 1825 he was buried on the mission grounds. The town of San Juan Bautista, which grew up around the mission, expanded rapidly during the California Gold Rush and continues to be a thriving community today.
The mission is situated adjacent to the San Andreas Fault, and has suffered damage from numerous earthquakes, such as those of 1800 and 1906. However, the mission was never entirely destroyed at once. It was restored initially in 1884, and then again in 1949 with funding from the Hearst Foundation. The three-bell campanario, or "bell wall," located by the church entrance, was fully restored in 2010. An unpaved stretch of the original El Camino Real, just east of the mission, lies on a fault scarp.
Although initially secularized in 1835, the church was reconsecrated by the Roman Catholic Church in 1859, and continues to serve as a parish of the Diocese of Monterey. The mission includes a cemetery, with the remains of over 4,000 Native American converts and Europeans buried there.
The mission and its grounds were featured prominently in the 1958 Alfred Hitchcock film Vertigo. Associate producer Herbert Coleman's daughter Judy Lanini suggested the mission to Hitchcock as a filming location. A steeple, added sometime after the mission's original construction and secularization, had been demolished following a fire, so Hitchcock added a bell tower using scale models, matte paintings, and trick photography at the Paramount studio in Los Angeles. The tower does not resemble the original steeple. The tower's staircase was assembled inside a studio.
See also
Spanish missions in California
List of Spanish missions in California
Rancho San Justo
Teatro Campesino
USNS Mission San Juan (AO-126) – a Buenaventura Class fleet oiler built during World War II.
References
Sources
External links
Mission San Juan Bautista – official site
Early photographs, sketches, land surveys of Mission San Juan Bautista, via Calisphere, California Digital Library
Listing and photographs of church at the Historic American Buildings Survey
Listing, drawings and photographs of mission at the Historic American Buildings Survey
Another view of the Mission Facade, circa 1980s
San Juan Bautista
San Juan Bautista, California
1797 in Alta California
Museums in San Benito County, California
Religious museums in California
History museums in California
Churches in San Benito County, California
Cemeteries in California
1797 establishments in Alta California
California Historical Landmarks
Churches on the National Register of Historic Places in California
Roman Catholic Diocese of Monterey in California
History of San Benito County, California
Spanish Colonial architecture in California
National Register of Historic Places in San Benito County, California |
null | null | German reunification | eng_Latn | German reunification () was the process in 1990 in which the German Democratic Republic (GDR; , DDR) became part of the Federal Republic of Germany (FRG; , BRD) to form the reunited nation of Germany.
The end of the unification process is officially referred to as German unity (), celebrated each year on 3 October as German Unity Day (). East and West Berlin were reunited into a single city, and again became the capital of united Germany.
The East German government started to falter in May 1989, when the removal of Hungary's border fence with Austria opened a hole in the Iron Curtain. The border was still closely guarded, but the Pan-European Picnic and the indecisive reaction of the rulers of the Eastern Bloc set in motion an irreversible peaceful movement. It allowed an exodus of thousands of East Germans fleeing to West Germany via Hungary. The Peaceful Revolution, a series of protests by East Germans, led to the GDR's first free elections on 18 March 1990, and to the negotiations between the GDR and FRG that culminated in a Unification Treaty. Other negotiations between the GDR and FRG and the four occupying powers produced the so-called "Two Plus Four Treaty" (Treaty on the Final Settlement with Respect to Germany) granting full sovereignty to a unified German state, whose two parts were previously bound by a number of limitations stemming from their post-World War II status as occupied regions.
The 1945 Potsdam Agreement had specified that a full peace treaty concluding World War II, including the exact delimitation of Germany's postwar boundaries, was required to be "accepted by the Government of Germany when a government adequate for the purpose is established." The Federal Republic had always maintained that no such government could be said to have been established until East and West Germany had been united within a free democratic state; but in 1990 a range of opinions continued to be maintained over whether a unified West Germany, East Germany, and Berlin could be said to represent "Germany as a whole" for this purpose. The key question was whether a Germany that remained bounded to the east by the Oder–Neisse line (the international border with Poland) could act as a "united Germany" in signing the peace treaty without qualification. Under the "Two Plus Four Treaty" both the Federal Republic and the Democratic Republic committed themselves and their unified continuation to the principle that their joint pre-1990 boundaries constituted the entire territory that could be claimed by a Government of Germany, and hence that there were no further lands outside those boundaries that were parts of Germany as a whole.
The post-1990 united Germany is not a successor state, but an enlarged continuation of the former West Germany. The enlarged Federal Republic of Germany retained the West German seats in international organizations including the European Economic Community (later the European Union), NATO, and the United Nations. Memberships in the Warsaw Pact and other international organizations to which East Germany belonged ended because East Germany ceased to exist.
Naming
For political and diplomatic reasons, West German politicians carefully avoided the term "reunification" during the run-up to what Germans frequently refer to as die Wende (roughly: the turning point). The 1990 treaty defines the official term as Deutsche Einheit ("German unity"); this is commonly used in Germany, and it is this term that then-Vice Chancellor Hans-Dietrich Genscher used in front of international journalists to correct them when they asked him about "reunification" in 1990.
After 1990, the term die Wende became more common. The term generally refers to the events (mostly in Eastern Europe) that led up to the actual reunification; in its usual context, this term loosely translates to "the turning point", without any further meaning. When referring to the events surrounding reunification, however, it carries the cultural connotation of the time and the events in the GDR that brought about this "turnaround" in German history. However, anti-communist activists from Eastern Germany rejected the term Wende as it had been introduced by the Socialist Unity Party of Germany's Secretary General Egon Krenz.
Precursors to reunification
Germany was officially divided into four occupation zones as result of the Potsdam Agreement on 1 August 1945, under the four military governments of the United States, the United Kingdom, France and the Soviet Union. The capital city of Berlin was similarly divided into the four sectors. Between 1947 and 1949, the three zones of the western allies were merged, forming the Federal Republic of Germany and West Berlin, aligned with capitalist Europe (which later developed into the European Community). The Soviet zone became the German Democratic Republic with its capital in East Berlin, a part of the communist Soviet Bloc. The FRG was a member of the western military alliance, NATO; the GDR was a member of the Warsaw Pact. Germans lived under such imposed divisions throughout the ensuing Cold War.
Into the 1980s, the Soviet Union experienced a period of economic and political stagnation, and correspondingly decreased intervention in Eastern Bloc politics. In 1987, US President Ronald Reagan gave a speech at the Brandenburg Gate, challenging Soviet General Secretary Mikhail Gorbachev to "tear down this wall" which divided Berlin. The wall had stood as an icon for the political and economic division between East and West, a division that Churchill had referred to as the "Iron Curtain". Gorbachev announced in 1988 that the Soviet Union would abandon the Brezhnev Doctrine and allow the Eastern bloc nations to freely determine their own internal affairs. In early 1989, under a new era of Soviet policies of glasnost (openness) and perestroika (economic restructuring), and taken further by Gorbachev, the Solidarity movement took hold in Poland. Further inspired by other images of brave defiance, a wave of revolutions swept throughout the Eastern Bloc that year.
In May 1989, Hungary removed their border fence. However, the dismantling of the old Hungarian border facilities did not open the borders, nor were the previous strict controls removed, and the isolation by the Iron Curtain was still intact over its entire length. The opening of a border gate between Austria and Hungary at the Pan-European Picnic on 19 August 1989 then set in motion a peaceful chain reaction, at the end of which there was no longer a GDR and the Eastern Bloc had disintegrated. Extensive advertising for the planned picnic was made by posters and flyers among the GDR holidaymakers in Hungary. The Austrian branch of the Paneuropean Union, which was then headed by Karl von Habsburg, distributed thousands of brochures inviting them to a picnic near the border at Sopron. It was the largest escape movement from East Germany since the Berlin Wall was built in 1961. After the picnic, which was based on an idea by Karl's father Otto von Habsburg to test the reaction of the USSR and Mikhail Gorbachev to an opening of the border, tens of thousands of media-informed East Germans set off for Hungary. The media reaction of Erich Honecker in the "Daily Mirror" of 19 August 1989 showed the public in East and West that there had been a loss of power by the Eastern European communist rulers in their own sphere of power, and that they were no longer the designers of what was happening: "Habsburg distributed leaflets far into Poland, on which the East German holidaymakers were invited to a picnic. When they came to the picnic, they were given gifts, food and Deutsche Mark, and then they were persuaded to come to the West." In particular, it was examined by Habsburg and the Hungarian Minister of State Imre Pozsgay, whether Moscow would give the Soviet troops stationed in Hungary the command to intervene. But with the mass exodus at the Pan-European Picnic, the subsequent hesitant behavior of the Socialist Unity Party of East Germany and the non-intervention of the Soviet Union broke the dams. Thus the bracket of the Eastern Bloc was broken.
Tens of thousands of the media-informed East Germans now made their way to Hungary, which was no longer ready to keep its borders completely closed or to oblige its border troops to use force of arms. By the end of September 1989, more than 30,000 East Germans had escaped to the West before the GDR denied travel to Hungary, leaving Czechoslovakia as the only neighboring state to which East Germans could escape.
Even then, many people within and without Germany still believed that real reunification would never happen in the foreseeable future. The turning point in Germany, called "Die Wende", was marked by the "Peaceful Revolution" leading to the removal of the Berlin Wall, with East and West Germany subsequently entering into negotiations toward eliminating the division that had been imposed upon Germans more than four decades earlier.
Process of reunification
Cooperation
On 28 November 1989—two weeks after the fall of the Berlin Wall—West German Chancellor Helmut Kohl announced a 10-point program calling for the two Germanies to expand their cooperation with a view toward eventual reunification.
Initially, no timetable was proposed. However, events rapidly came to a head in early 1990. First, in March, the Party of Democratic Socialism—the former Socialist Unity Party of Germany—was heavily defeated in East Germany's first free elections. A grand coalition was formed under Lothar de Maizière, leader of the East German wing of Kohl's Christian Democratic Union, on a platform of speedy reunification. Second, East Germany's economy and infrastructure underwent a swift and near-total collapse. Although East Germany was long reckoned as having the most robust economy in the Soviet bloc, the removal of Communist hegemony revealed the ramshackle foundations of that system. The East German mark had been almost worthless outside East Germany for some time before the events of 1989–90, and the collapse of the East German economy further magnified the problem.
Economic merger
Discussions immediately began on an emergency merger of the German economies. On 18 May 1990, the two German states signed a treaty agreeing on monetary, economic and social union. This treaty is called ("Treaty Establishing a Monetary, Economic and Social Union between the German Democratic Republic and the Federal Republic of Germany"); it came into force on 1 July 1990, with the West German Deutsche Mark replacing the East German mark as the official currency of East Germany. The Deutsche Mark had a very high reputation among the East Germans and was considered stable. While the GDR transferred its financial policy sovereignty to West Germany, the West started granting subsidies for the GDR budget and social security system. At the same time, many West German laws came into force in the GDR. This created a suitable framework for a political union by diminishing the huge gap between the two existing political, social, and economic systems.
German Reunification Treaty
The Volkskammer, the Parliament of East Germany, passed a resolution on 23 August 1990 declaring the accession (Beitritt) of the German Democratic Republic to the Federal Republic of Germany, and the extension of the field of application of the Federal Republic's Basic Law to the territory of East Germany as allowed by article 23 of the West German Basic Law, effective 3 October 1990. This Declaration of Accession (Beitrittserklärung) was formally presented by the President of the Volkskammer, Sabine Bergmann-Pohl, to the President of the West German Bundestag, Rita Süssmuth, by means of a letter dated 25 August 1990. Thus, formally, the procedure of reunification by means of the accession of East Germany to West Germany, and of East Germany's acceptance of the Basic Law already in force in West Germany, was initiated as the unilateral, sovereign decision of East Germany, as allowed by the provisions of article 23 of the West German Basic Law as it then existed.
In the wake of that resolution of accession, the "German reunification treaty", commonly known in German as "Einigungsvertrag" (Unification Treaty) or "Wiedervereinigungsvertrag" (Reunification Treaty), that had been negotiated between the two German states since 2 July 1990, was signed by representatives of the two Governments on 31 August 1990. This Treaty, officially titled Vertrag zwischen der Bundesrepublik Deutschland und der Deutschen Demokratischen Republik über die Herstellung der Einheit Deutschlands (Treaty between the Federal Republic of Germany and the German Democratic Republic on the Establishment of German Unity), was approved by large majorities in the legislative chambers of both countries on 20 September 1990 (442–47 in the West German Bundestag and 299–80 in the East German Volkskammer). The Treaty passed the West German Bundesrat on the following day, 21 September 1990. The amendments to the Federal Republic's Basic Law that were foreseen in the Unification Treaty or necessary for its implementation were adopted by the Federal Statute of 23 September 1990, that enacted the incorporation of the Treaty as part of the Law of the Federal Republic of Germany. The said Federal Statute, containing the whole text of the Treaty and its Protocols as an annex, was published in the Bundesgesetzblatt (the official journal for the publication of the laws of the Federal Republic) on 28 September 1990. In the German Democratic Republic, the constitutional law (Verfassungsgesetz) giving effect to the Treaty was also published on 28 September 1990. With the adoption of the Treaty as part of its Constitution, East Germany legislated its own abolition as a State.
Under article 45 of the Treaty, it entered into force according to international law on 29 September 1990, upon the exchange of notices regarding the completion of the respective internal constitutional requirements for the adoption of the treaty in both East Germany and West Germany. With that last step, and in accordance with article 1 of the Treaty, and in conformity with East Germany's Declaration of Accession presented to the Federal Republic, Germany was officially reunited at 00:00 CEST on 3 October 1990. East Germany joined the Federal Republic as the five Länder (states) of Brandenburg, Mecklenburg-Western Pomerania, Saxony, Saxony-Anhalt and Thuringia. These states were the five original states of East Germany, but were abolished in 1952 in favor of a centralized system. As part of 18 May treaty, the five East German states were reconstituted on 23 August. At the same time, East and West Berlin reunited into one city, which became a city-state along the lines of the existing city-states of Bremen and Hamburg. Berlin was still formally under Allied occupation (that would only be terminated later, as a result of the provisions of the Two Plus Four Treaty), but the city's administrative merger and inclusion in the Federal Republic of Germany, effective on 3 October 1990, had been greenlighted by the Allies, and were formally approved in the final meeting of the Allied Control Council on 2 October 1990. In an emotional ceremony, at the stroke of midnight on 3 October 1990, the black-red-gold flag of West Germany—now the flag of a reunited Germany—was raised above the Brandenburg Gate marking the moment of German reunification.
Constitutional merger
The process chosen was one of two options implemented in the West German constitution (Grundgesetz or Basic Law) of 1949 to facilitate eventual reunification. The Basic Law stated that it was only intended for temporary use until a permanent constitution could be adopted by the German people as a whole. Via that document's (then-existing) Article 23, any new prospective Länder could adhere to the Basic Law by a simple majority vote. The initial eleven joining states of 1949 constituted the Trizone. West Berlin had been proposed as the 12th state, but was legally inhibited by Allied objections since Berlin as a whole was legally a quadripartite occupied area. Despite this, West Berlin's political affiliation was with West Germany, and in many fields, it functioned de facto as if it were a component state of West Germany. In 1957 the Saar Protectorate joined West Germany under the Article 23 procedure as Saarland.
The other option was Article 146, which provided a mechanism for a permanent constitution for a reunified Germany. This route would have entailed a formal union between two German states that then would have had to, amongst other things, create a new constitution for the newly established country. However, by the spring of 1990, it was apparent that drafting a new constitution would require protracted negotiations that would open up numerous issues in West Germany. Even without this to consider, by the start of 1990 East Germany was in a state of economic and political collapse. In contrast, reunification under Article 23 could be implemented in as little as six months.
Ultimately, when the treaty on monetary, economic and social union was signed, it was decided to use the quicker process of Article 23. By this process, East Germany voted to dissolve itself and to join West Germany, and the area in which the Basic Law was in force simply extended to include them. Thus, while legally East Germany as a whole acceded to the Federal Republic, the constituent parts of East Germany entered into the Federal Republic as five new states, which held their first elections on 14 October 1990.
Nevertheless, although the Volkskammer's declaration of accession to the Federal Republic had initiated the process of reunification, the act of reunification itself (with its many specific terms, conditions, and qualifications, some of which required amendments to the Basic Law itself) was achieved constitutionally by the subsequent Unification Treaty of 31 August 1990; that is through a binding agreement between the former GDR and the Federal Republic now recognising each another as separate sovereign states in international law. This treaty was then voted into effect by both the Volkskammer and the Bundestag by the constitutionally required two-thirds majorities, effecting on the one hand, the extinction of the GDR, and on the other, the agreed amendments to the Basic Law of the Federal Republic. Hence, although the GDR declared its accession to the Federal Republic under Article 23 of the Basic Law, this did not imply its acceptance of the Basic Law as it then stood, but rather, of the Basic Law as subsequently amended in line with the Unification Treaty.
Legally, the reunification did not create a third state out of the two. Rather, West Germany effectively absorbed East Germany. Accordingly, on Unification Day, 3 October 1990, the German Democratic Republic ceased to exist, and five new federated states on its former territory joined the Federal Republic of Germany. East and West Berlin were reunited and joined the Federal Republic as a full-fledged federated city-state. Under this model, the Federal Republic of Germany, now enlarged to include the five states of the former GDR plus the reunified Berlin, continued legally to exist under the same legal personality that was founded in May 1949.
While the Basic Law was modified, rather than replaced by a constitution as such, it still permits the adoption of a formal constitution by the German people at some time in the future.
International effects
The practical result of that model is that the now-expanded Federal Republic of Germany inherited the old West Germany's seats at the UN, NATO, the European Communities and other international organizations. It also continued to be a party to all the treaties the old West Germany signed prior to the moment of reunification. The Basic Law and statutory laws that were in force in the Federal Republic, as amended in accordance with the Unification Treaty, continued automatically in force, but now applied to the expanded territory. Also, the same President, Chancellor (Prime Minister) and Government of the Federal Republic remained in office, but their jurisdiction now included the newly acquired territory of the former East Germany.
To facilitate this process and to reassure other countries, fundamental changes were made to the "Basic Law" (constitution). The Preamble and Article 146 were amended, and Article 23 was replaced, but the deleted former Article 23 was applied as the constitutional model to be used for the 1990 reunification. Hence, prior to the five "New Länder" of East Germany joining, the Basic Law was amended to indicate that all parts of Germany would then be unified such that Germany could now no longer consider itself constitutionally open to further extension to include the former eastern territories of Germany, that were now Polish, Russian or Lithuanian. The changes effectively formalized the Oder–Neisse line as Germany's permanent eastern border. These amendments to the Basic Law were mandated by Article I, section 4 of the Two Plus Four Treaty.
Day of German Unity
To commemorate the day that marks the official unification of the former East and West Germany in 1990, 3 October has since then been the official German national holiday, the Day of German Unity (Tag der deutschen Einheit). It replaced the previous national holiday held in West Germany on 17 June commemorating the Uprising of 1953 in East Germany and the national holiday on 7 October in the GDR, that commemorated the foundation of the East German state.
Foreign support and opposition
For decades, West Germany's allies stated their support for reunification. Israeli Prime Minister Yitzhak Shamir, who speculated that a country that "decided to kill millions of Jewish people" in the Holocaust "will try to do it again", was one of the few world leaders to publicly oppose it. As reunification became a realistic possibility, however, significant NATO and European opposition emerged in private.
A poll of four countries in January 1990 found that a majority of surveyed Americans and French supported reunification, while British and Poles were more divided. 69% of Poles and 50% of French and British stated that they worried about a reunified Germany becoming "the dominant power in Europe". Those surveyed stated several concerns, including Germany again attempting to expand its territory, a revival of Nazism, and the German economy becoming too powerful. While British, French, and Americans favored Germany remaining a member of NATO, a majority of Poles supported neutrality for the reunified nation.
The key ally was the United States. Although some top American officials opposed quick unification, Secretary of State James A. Baker and President George H. W. Bush provided strong and decisive support to Kohl's proposals.
Britain and France
Before the fall of the Berlin Wall, British Prime Minister Margaret Thatcher told Soviet General Secretary Mikhail Gorbachev that neither the United Kingdom nor Western Europe wanted the reunification of Germany. Thatcher also clarified she wanted the Soviet leader to do what he could to stop it, telling Gorbachev "We do not want a united Germany". Although she welcomed East German democracy, Thatcher worried that a rapid reunification might weaken Gorbachev, and favored Soviet troops staying in East Germany as long as possible to act as a counterweight to a united Germany.
Thatcher, who carried in her handbag a map of Germany's 1937 borders to show others the "German problem", feared that its "national character", size and central location in Europe would cause the nation to be a "destabilizing rather than a stabilizing force in Europe". In December 1989, she warned fellow European Community leaders at a Strasbourg summit that Kohl attended, "We defeated the Germans twice! And now they're back!" Although Thatcher had stated her support for German self-determination in 1985, she now argued that Germany's allies only supported reunification because they did not believe it would ever happen. Thatcher favored a transition period of five years for reunification, during which the two Germanies would remain separate states. Although she gradually softened her opposition, as late as March 1990 Thatcher summoned historians and diplomats to a seminar at Chequers to ask "How dangerous are the Germans?" and the French ambassador in London reported that Thatcher told him, "France and Great Britain should pull together today in the face of the German threat."
The pace of events surprised the French, whose Foreign Ministry had concluded in October 1989 that reunification "does not appear realistic at this moment". A representative of French President François Mitterrand reportedly told an aide to Gorbachev, "France by no means wants German reunification, although it realises that in the end, it is inevitable." At the Strasbourg summit, Mitterrand and Thatcher discussed the fluidity of Germany's historical borders. On 20 January 1990, Mitterrand told Thatcher that a unified Germany could "make more ground than even Adolf had". He predicted that "bad" Germans would reemerge, who might seek to regain former German territory lost after World War II and would likely dominate Hungary, Poland, and Czechoslovakia, leaving "only Romania and Bulgaria for the rest of us". The two leaders saw no way to prevent reunification, however, as "None of us was going to declare war on Germany". Mitterrand recognized before Thatcher that reunification was inevitable and adjusted his views accordingly; unlike her, he was hopeful that participation in a single currency and other European institutions could control a united Germany. Mitterrand still wanted Thatcher to publicly oppose unification, however, to obtain more concessions from Germany.
Rest of Europe
Ireland's Taoiseach, Charles Haughey, supported German reunification and he took advantage of Ireland's presidency of the European Economic Community to call for an extraordinary European summit in Dublin in April 1990 to calm the fears held of fellow members of the EEC. Haughey saw similarities between Ireland and Germany, and said "I have expressed a personal view that coming as we do from a country which is also divided many of us would have sympathy with any wish of the people of the two German States for unification". Der Spiegel later described other European leaders' opinion of reunification at the time as "icy". Italy's Giulio Andreotti warned against a revival of "pan-Germanism" and joked "I love Germany so much that I prefer to see two of them", and the Netherlands' Ruud Lubbers questioned the German right to self-determination. They shared Britain and France's concerns over a return to German militarism and the economic power of a reunified nation. The consensus opinion was that reunification, if it must occur, should not occur until at least 1995 and preferably much later.
Four powers
The victors of World War II — France, the Soviet Union, the United Kingdom, and the United States, comprising the Four-Power Authorities—retained authority over Berlin, such as control over air travel and its political status. From the onset, the Soviet Union sought to use reunification as a way to push Germany out of NATO into neutrality, removing nuclear weapons from its territory. However, West Germany misinterpreted a 21 November 1989 diplomatic message on the topic to mean that the Soviet leadership already anticipated reunification only two weeks after the Wall's collapse. This belief, and the worry that his rival Genscher might act first, encouraged Kohl on 28 November to announce a detailed "Ten Point Program for Overcoming the Division of Germany and Europe". While his speech was very popular within West Germany, it caused concern among other European governments, with whom he had not discussed the plan.
The Americans did not share the Europeans' and Soviets' historical fears over German expansionism; Condoleezza Rice later recalled,
The United States wished to ensure, however, that Germany would stay within NATO. In December 1989, the administration of President George H. W. Bush made a united Germany's continued NATO membership a requirement for supporting reunification. Kohl agreed, although less than 20% of West Germans supported remaining within NATO. Kohl also wished to avoid a neutral Germany, as he believed that would destroy NATO, cause the United States and Canada to leave Europe, and cause Britain and France to form an anti-German alliance. The United States increased its support of Kohl's policies, as it feared that otherwise Oskar Lafontaine, a critic of NATO, might become Chancellor.
Horst Teltschik, Kohl's foreign policy advisor, later recalled that Germany would have paid "100 billion deutschmarks" if the Soviets demanded it. The USSR did not make such great demands, however, with Gorbachev stating in February 1990 that "The Germans must decide for themselves what path they choose to follow". In May 1990 he repeated his remark in the context of NATO membership while meeting Bush, amazing both the Americans and Germans. This removed the last significant roadblock to Germany being free to choose its international alignments, though Kohl made no secret that he intended for the reunified Germany to inherit West Germany's seats in NATO and the EC.
Conclusion
During a NATO–Warsaw Pact conference in Ottawa, Canada, Genscher persuaded the four powers to treat the two Germanies as equals instead of defeated junior partners, and for the six nations to negotiate alone. Although the Dutch, Italians, Spanish, and other NATO powers opposed such a structure, which meant that the alliance's boundaries would change without their participation, the six nations began negotiations in March 1990. After Gorbachev's May agreement on German NATO membership, the Soviets further agreed that Germany would be treated as an ordinary NATO country, with the exception that former East German territory would not have foreign NATO troops or nuclear weapons. In exchange, Kohl agreed to reduce the sizes of the militaries of both West and East Germany, renounce weapons of mass destruction, and accept the postwar Oder–Neisse line as Germany's eastern border. In addition, Germany agreed to pay about 55 billion deutschmarks to the Soviet Union in gifts and loans, the equivalent of eight days of the West German GDP.
The British insisted to the end, against Soviet opposition, that NATO be allowed to hold manoeuvres in the former East Germany. After the Americans intervened, both the UK and France ratified the Treaty on the Final Settlement with Respect to Germany in September 1990, thus finalizing the reunification for purposes of international law. Thatcher later wrote that her opposition to reunification had been an "unambiguous failure".
Aftermath
German sovereignty, confirmation of borders, withdrawal of the Allied Forces
On 14 November 1990, Germany and Poland signed the German–Polish Border Treaty, finalizing Germany's boundaries as permanent along the Oder–Neisse line, and thus, renouncing any claims to Silesia, East Brandenburg, Farther Pomerania, and the southern area of the former province of East Prussia. The subsequent German–Polish Treaty of Good Neighbourship that supplemented the Border Treaty also granted certain rights for political minorities on either side of the border. The following month, the first all-German free elections since 1932 were held, resulting in an increased majority for the coalition government of Chancellor Helmut Kohl.
On 15 March 1991, the Treaty on the Final Settlement with Respect to Germany—that had been signed in Moscow back on 12 September 1990 by the two German states that then existed (East and West Germany) on one side, and by the four principal Allied powers (the United Kingdom, France, the Soviet Union and the United States) on the other—entered into force, having been ratified by the Federal Republic of Germany (after the unification, as the united Germany) and by the four Allied nations. The entry into force of that treaty (also known as the "Two Plus Four Treaty", in reference to the two German states and four Allied nations that signed it) put an end to the then-remaining limitations on German sovereignty that resulted from the post-World War II arrangements.
Even prior to the ratification of the Treaty, the operation of all quadripartite Allied institutions in Germany was suspended, with effect from the reunification of Germany on 3 October 1990 and pending the final ratification of the Two Plus Four Treaty, pursuant to a declaration signed in New York on 1 October 1990 by the foreign ministers of the four Allied Powers, that was witnessed by ministers of the two German states then in existence, and that was appended text of the Two Plus Four Treaty.
Article 5 banned the deployment of nuclear weapons in the territory previously controlled by the DDR and well as a ban on stationing non-German military personnel.
In accordance with Article 9 of the Two Plus Four Treaty, it entered into force as soon as all ratifications were deposited with the Government of Germany. The last party to ratify the treaty was the Soviet Union, that deposited its instrument of ratification on 15 March 1991. The Supreme Soviet of the USSR only gave its approval to the ratification of the treaty on 4 March 1991, after a hefty debate.
Under that treaty (which should not be confused with the Unification Treaty that was signed only between the two German states), the last Allied forces still present in Germany left in 1994, in accordance with article 4 of the treaty, that set 31 December 1994 as the deadline for the withdrawal of the remaining Allied forces. The bulk of Russian ground forces left Germany on 25 June 1994 with a military parade of the 6th Guards Motor Rifle Brigade in Berlin. This was followed by the closure of the United States Army Berlin command on 12 July 1994, an event that was marked by a casing of the colours ceremony witnessed by President Bill Clinton. The withdrawal of the last Russian troops (the Russian Army's Western Group of Forces) was completed on 31 August 1994, and the event was marked by a military ceremony in the Treptow Park in Berlin, with the presence of Russian President Yeltsin and German Chancellor Kohl. Although the bulk of the British, American, and French Forces had left Germany even before the departure of the Russians, the Western Allies kept a presence in Berlin until the completion of the Russian withdrawal, and the ceremony marking the departure of the remaining Forces of the Western Allies was the last to take place: on 8 September 1994, a Farewell Ceremony in the courtyard of the Charlottenburg Palace, with the presence of British Prime Minister John Major, American Secretary of State Warren Christopher, French President François Mitterrand, and German Chancellor Helmut Kohl, marked the withdrawal of the British, American and French Occupation Forces from Berlin, and the termination of the Allied occupation in Germany. Thus, the removal of the Allied presence took place a few months before the final deadline.
As for the German–Polish Border Treaty, it was approved by the Polish Sejm on 26 November 1991 and the German Bundestag on 16 December 1991, and entered into force with the exchange of the instruments of ratification on 16 January 1992. The confirmation of the border between Germany and Poland was required of Germany by the Allied Powers in the Two Plus Four Treaty.
Cost of reunification
The subsequent economic restructuring and reconstruction of eastern Germany resulted in significant costs, especially for western Germany, which paid large sums of money in the form of the Solidaritätszuschlag (Solidarity Surcharge) in order to rebuild the east German infrastructure. In addition, the immensely advantageous exchange rate of 1:1 between the West German Deustchmark to the East German mark meant that East Germans could trade in their almost worthless marks for and receive wages in West German currency. This dealt a major blow to the West German budget in the coming few years. Peer Steinbrück is quoted as saying in a 2011 interview, "Over a period of 20 years, German reunification has cost 2 trillion euros, or an average of 100 billion euros a year. So, we have to ask ourselves 'Aren't we willing to pay a tenth of that over several years for Europe's unity?'"
Views and life satisfaction
According to a 2019 survey conducted by Pew Research Center, approximately 90 percent of Germans living in both the West and East believe that reunification was good for Germany, with slightly more in East than West Germany supporting it. Around 83 percent of East Germans approve of and 13 percent disapprove of eastern Germany's transition to a market economy, with the rest saying they weren't sure. Life satisfaction in both the East and West has substantially increased since 1991, with 15 percent of East Germans placing their life satisfaction somewhere between 7 to 10 on a 0 to 10 scale in 1991 changing to 59 percent in 2019. For West Germans, this change over the same time period was from 52 to 64 percent.
Additionally, the fall of the Berlin Wall was useful in generating wealth at the household level in both the East and the West. Those who lived in West Germany and had social ties to the East experienced a six percent average increase in their wealth in the six years following the fall of the wall, which more than doubled that of households who didn't possess the same connections. Entrepreneurs who worked in areas with strong social ties to the East saw their incomes increase at an even more rapid rate. Incomes for this group increased at an average rate of 8.8 percent over the same six year period following reunification. Similarly, those in the East who possessed connections to the West saw their household income increase at a positive rate in each of the six years following reunification. Those in their regions who lacked the same ties did not see the same benefit. More broadly, commercial firms, in the same way as households and individuals, also saw profits increase in the years following reunification. Specifically, gains in the form of increased profits were the greatest for firms in the service sector who invested in the East relative to those which did not. The reunification served to increase economic growth in the region through rising household income and commercial profit due to the ability to utilize social connections that were previously restricted by the border.
Inner reunification
Vast differences between former East Germany and West Germany in lifestyle, wealth, political beliefs, and other matters remain, and it is therefore still common to speak of eastern and western Germany distinctly. It is often referred to as the "wall in the head" ("Mauer im Kopf"). Ossis (Easterners) are stereotyped as racist, poor and largely influenced by Russian culture, while Wessis (Westerners) are usually considered snobbish, dishonest, wealthy, and selfish. East Germans indicate a dissatisfaction with the status quo and cultural alienation from the rest of Germany, and a sense that their cultural heritage is not acknowledged enough in the now unified Germany. The West, on the other hand, has become uninterested in what the East has to say, and this has led to more resentment toward the East exacerbating the divide. Both the West and the East have failed to sustain an open-minded dialogue and the failure to grasp the effects of the institutional path dependency has increased the frustration each side feels.
The economy of eastern Germany has struggled since unification, and large subsidies are still transferred from west to east. Economically, eastern Germany has had a sharp rise of 10% to West Germany’s 5%. Western Germany also still holds 56% of the GDP. Part of this disparity between the East and the West lies in the Western labor Unions' demand for high-wage pacts in an attempt to prevent "low-wage zones." This caused many Germans from the East to be outpriced in the market, adding to the slump in businesses in eastern Germany as well as the rising unemployment. The former East German area has often been compared to the underdeveloped Southern Italy and the Southern United States during Reconstruction after the American Civil War. While the economy of eastern Germany has recovered recently, the differences between East and West remain present.
Politicians and scholars have frequently called for a process of "inner reunification" of the two countries and asked whether there is "inner unification or continued separation". "The process of German unity has not ended yet", proclaimed Chancellor Angela Merkel, who grew up in East Germany, in 2009. Nevertheless, the question of this "inner reunification" has been widely discussed in the German public, politically, economically, culturally, and also constitutionally since 1989.
Politically, since the fall of the Berlin Wall, the successor party of the former East German socialist state party has become a major force in German politics. It was renamed PDS, and, later, merged with the Western leftist party WASG to form the party The Left (Die Linke).
Constitutionally, the Basic Law of West Germany (Grundgesetz) provided two pathways for unification. The first was the implementation of a new all-German constitution, safeguarded by a popular referendum. Actually, this was the original idea of the "Grundgesetz" in 1949: it was named a "basic law" instead of a "constitution" because it was considered provisional. The second way was more technical: the implementation of the constitution in the East, using a paragraph originally designed for the West German states (Bundesländer) in case of internal re-organization like the merger of two states. While this latter option was chosen as the most feasible one, the first option was partly regarded as a means to foster the "inner reunification".
A public manifestation of coming to terms with the past (Vergangenheitsbewältigung) is the existence of the so-called Birthler-Behörde, the Federal Commissioner for the Stasi Records, which collects and maintains the files of the East German security apparatus.
The economic reconstruction of former East Germany following the reunification required large amounts of public funding which turned some areas into boom regions, although overall unemployment remains higher than in the former West.
Unemployment was part of a process of deindustrialization starting rapidly after 1990. Causes for this process are disputed in political conflicts up to the present day. Most times bureaucracy and lack of efficiency of the East German economy are highlighted and the de-industrialization is seen as an inevitable outcome of the "Wende". But many critics from East Germany point out that it was the shock-therapy style of privatization that did not leave room for East German enterprises to adapt, and that alternatives like a slow transition had been possible.
Reunification did, however, lead to a large rise in the average standard of living in former East Germany and a stagnation in the West as $2 trillion in public spending was transferred East. Between 1990 and 1995, gross wages in the east rose from 35% to 74% of western levels, while pensions rose from 40% to 79%. Unemployment reached double the western level as well. West German cities close to the former border of East and West Germany experienced a disproportionate loss of market access relative to other West German cities which were not as greatly affected by the reunification of Germany.
In terms of media usage and reception, the country remains partially divided especially among the older generations. Mentality gaps between East and West persist, but so does sympathy. Additionally, the integration between Easterners and Westerners is not happening on as large a scale as was expected. Young people have on average very little knowledge of former East Germany. Some people in Eastern Germany engage in "Ostalgie", which is a certain nostalgia for the time before the wall came down.
Today, there are several prominent people of East German origin, including Michael Ballack, Katarina Witt, Paul van Dyk, and Angela Merkel.
Reunified Berlin
While the fall of the Berlin Wall had broad economic, political and social impacts globally, it also had significant consequence for the local urban environment. In fact, the events of 9 November 1989 saw East Berlin and West Berlin, two halves of a single city that had ignored one another for the better part of 40 years, finally "in confrontation with one another". There was a belief in the city that after 40 years of division, the reunified city would be well placed to become a major metropolis.
In the context of urban planning, in addition to a wealth of new opportunity and the symbolism of two former independent nations being re-joined, the reunification of Berlin presented numerous challenges. The city underwent massive redevelopment, involving the political, economic, and cultural environment of both East and West Berlin. However, the "scar" left by the Wall, which ran directly through the very heart of the city had consequences for the urban environment that planning still needs to address.
Urban planning issues
The reunification of Berlin presented legal, political, and technical challenges for the urban environment. The political division and physical separation of the city for more than 30 years saw the East and the West develop their own distinct urban forms, with many of these differences still visible to this day.
East and West Berlin were directed by two separate political and urban agendas. East Berlin developed a mono-centric structure around government buildings and open spaces, while West Berlin was poly-centric in nature, with a center that was higher-density but less residential. The two political systems allocated funds to post-war reconstruction differently, based on political priorities, and this had consequences for the reunification of the city. West Berlin had received considerably more financial assistance for reconstruction and refurbishment. There was considerable disparity in the general condition of many of the buildings: at the time of reunification, East Berlin still contained many leveled areas, which were previous sites of destroyed buildings from World War II, as well as damaged buildings that had not been repaired.
An immediate challenge facing the reunified city was the need for physical connectivity between the East and the West, specifically the organization of infrastructure. In the period following World War II, approximately half of the railway lines were removed in East Berlin. Urban rail required substantial work over more than a decade to fully reconnect the two halves of the city, and the tram network had been removed from the West leaving it entirely in the East.
Policy for reunification
As urban planning in Germany is the responsibility of the city government, the integration of East and West Berlin was in part complicated by the fact that the existing planning frameworks became obsolete with the fall of the Wall.
Prior to the reunification of the city, the Land Use Plan of 1988 and General Development Plan of 1980 defined the spatial planning criteria for West and East Berlin, respectively. These were replaced by the new, unified Land Use Plan in 1994. Termed "Critical Reconstruction", the new policy aimed to revive Berlin's pre-war aesthetic; it was complemented by a strategic planning document for downtown Berlin, entitled "Inner City Planning Framework".
Following the dissolution of the German Democratic Republic on 3 October 1990, all planning projects under the socialist regime were abandoned.
Vacant lots, open areas, and empty fields in East Berlin were subject to redevelopment, in addition to space previously occupied by the Wall and associated buffer zone. Many of these sites were positioned in central, strategic locations of the reunified city.
After the fall of the Wall
Berlin's urban organization experienced significant upheaval following the physical and metaphorical collapse of the Wall, as the city sought to "re-invent itself as a 'Western' metropolis".
Redevelopment of vacant lots, open areas, and empty fields, as well as space previously occupied by the Wall and associated buffer zone, were based on land use priorities as reflected in "Critical Reconstruction" policies. Green space and recreational areas were allocated 38% of freed land; 6% of freed land was dedicated to mass-transit systems to address transport inadequacies.
Reunification initiatives also included the construction of major office and commercial projects, as well as the renovation of housing estates in East Berlin.
Another key priority was reestablishing Berlin as the seat of government of Germany, and this required buildings to serve government needs, including the "redevelopment of sites for scores of foreign embassies".
With respect to redefining the city's identity, emphasis was placed on restoring Berlin's traditional landscape. "Critical Reconstruction" policies sought to disassociate the city's identity from its Nazi and socialist legacy, though some remnants were preserved, with walkways and bicycle paths established along the border strip to preserve the memory of the Wall. In the center of East Berlin much of the modernist heritage of the East German state was gradually removed. Reunification saw the removal of politically motivated street names and monuments in the East in an attempt to reduce the socialist legacy from the face of East Berlin.
Immediately following the fall of the Wall, Berlin experienced a boom in the construction industry. Redevelopment initiatives saw Berlin turn into one of the largest construction sites in the world through the 1990s and early 2000s.
The fall of the Wall also had economic consequences. Two German systems covering distinctly divergent degrees of economic opportunity suddenly came into intimate contact. Despite development of sites for commercial purposes, Berlin struggled to compete in economic terms with key West German centers such as Stuttgart and Düsseldorf. The intensive building activity directed by planning policy resulted in the over-expansion of office space, "with a high level of vacancies in spite of the move of most administrations and government agencies from Bonn".
Berlin was marred by disjointed economic restructuring, associated with massive deindustrialization. Economist Hartwich asserts that while the East undoubtedly improved economically, it was "at a much slower pace than [then Chancellor Helmut] Kohl had predicted". Wealth and income inequality between former East and West Germany continued for decades after reunification. On average adults in the former West Germany had assets worth 94,000 euros in 2014 as compared to the adults in the former communist East Germany which had just over 40,000 euros in assets.
Facilitation of economic development through planning measures failed to close the disparity between East and West, not only in terms of the economic opportunity but also housing conditions and transport options. Tölle states that "the initial euphoria about having become one unified people again was increasingly replaced by a growing sense of difference between Easterners ("Ossis") and Westerners ("Wessis")". The fall of the Wall also instigated immediate cultural change. The first consequence was the closure in East Berlin of politically oriented cultural institutions.
The fall of the Berlin Wall and the factors described above led to mass migration from East Berlin and East Germany, producing a large labor supply shock in the West. Emigration from the East, totaling 870,000 people between 1989 and 1992 alone, led to worse employment outcomes for the least-educated workers, for blue-collar workers, for men and for foreign nationals.
At the close of the century, it became evident that despite significant investment and planning, Berlin was yet to retake "its seat between the European Global Cities of London and Paris." Yet ultimately, the disparity between East and West portions of Berlin has led to the city achieving a new urban identity.
A number of locales of East Berlin, characterized by dwellings of in-between use of abandoned space for little to no rent, have become the focal point and foundation of Berlin's burgeoning creative activities. According to Berlin Mayor Klaus Wowereit, "the best that Berlin has to offer, its unique creativity. Creativity is Berlin's future." Overall, the Berlin government's engagement in creativity is strongly centered on marketing and promotional initiatives instead of creative production.
Creativity has been the catalyst for the city's "thriving music scene, active nightlife, and bustling street scene" all of which have become important attractions for the German capital. The industry is a key component of the city's economic make-up with more than 10% of all Berlin residents employed in cultural sectors.
Comparison
Germany was not the only state that had been separated through the aftermaths of World War II. For example, Korea (1945-present) as well as Vietnam (1954-1976) have been separated through the occupation of "Western-Capitalistic" and "Eastern-Communistic" forces, after the defeat of the Empire of Japan. Both countries suffered severely from this separation in the Korean War (1950–1953) and the Vietnam War (1955–1975) respectively, which caused heavy economic and civilian damage. However, German separation did not result in another war. Moreover, Germany is the only one of these countries that has managed to achieve a peaceful reunification. For instance, Vietnam achieved reunification after Vietnam War in 1976, while North and South Korea still struggle with high political tensions and huge economic and social disparities, making a possible reunification an enormous challenge.
See also
History of Germany (1945-1990)
Reunification
Stalin Note – 1952 German reunification proposal
Transitology
Good Bye, Lenin!
Chinese unification
Cypriot reunification
Irish reunification
Korean reunification
Yemeni unification
Unification of Romania and Moldova
Vietnamese reunification
References
Further reading
Blumenau, Bernhard, 'German Foreign Policy and the ‘German Problem’ During and After the Cold War: Changes and Continuities'. in: B Blumenau, J Hanhimäki & B Zanchetta (eds), New Perspectives on the End of the Cold War: Unexpected Transformations? Ch. 5. London: Routledge, 2018. .
Engel, Jeffrey A. When the World Seemed New: George H. W. Bush and the End of the Cold War (2018) pp. 273–291.
Maier, Charles S., Dissolution: The Crisis of Communism and the End of East Germany (Princeton UP, 1997).
Meacham, Jon. Destiny and Power: The American Odyssey of George Herbert Walker Bush (2015), pp. 397–408.
Schemper, Lukas. "Diasporas and American debates on German unification." Journal of Transatlantic Studies 15.1 (2017): 41-60 online.
Spohr, Kristina. "German Unification: Between Official History, Academic Scholarship, and Political Memoirs" Historical Journal 43#3 (2000), pp. 869–888, at p. 876. online.
Zelikow, Philip and Condoleezza Rice, Germany Unified and Europe Transformed: A Study in Statecraft (Harvard University Press, 1997) excerpt.
Primary sources
Jarausch, Konrad H., and Volker Gransow, eds. Uniting Germany: Documents and Debates, 1944–1993 (1994), primary sources in English translation
External links
The Unification Treaty (Berlin, 31 August 1990) website of CVCE (Centre of European Studies)
Hessler, Uwe, "The End of East Germany", dw-world.de, 23 August 2005.
Berg, Stefan, Steffen Winter and Andreas Wassermann, "Germany's Eastern Burden: The Price of a Failed Reunification", Der Spiegel, 5 September 2005.
Wiegrefe, Klaus, "An Inside Look at the Reunification Negotiations"Der Spiegel, 29 September 2010.
"Unfriendly, even dangerous"? Margaret Thatcher and German Unification, Academia.edu, 2016.
Problems with Reunification from the Dean Peter Krogh Foreign Affairs Digital Archives
Contemporary German history
Reunification
1990 in East Germany
1990 in West Germany
1990 in international relations
National unifications
Germany–Soviet Union relations
Peaceful Revolution
Ostalgie |
null | null | Raymond Burr | eng_Latn | William Raymond Stacy Burr (May 21, 1917September 12, 1993) was a Canadian-American actor known for his lengthy Hollywood film career and his title roles in the television dramas Perry Mason and Ironside.
Burr's early acting career included roles on Broadway, radio, television, and film, usually as the villain. His portrayal of the suspected murderer in the Alfred Hitchcock thriller Rear Window (1954) is his best-known film role, although he is also remembered for his role in the 1956 film Godzilla, King of the Monsters!, which he reprised in the 1985 film Godzilla 1985. He won Emmy Awards for acting in 1959 and 1961 for the role of Perry Mason, which he played for nine seasons (1957–1966) and reprised in a series of 26 Perry Mason TV movies (1985–1993). His second TV series Ironside earned him six Emmy and two Golden Globe nominations.
Burr died of cancer in 1993, and his personal life came into question, as many details of his biography appeared to be unverifiable. He was ranked number 44 of the 50 Greatest TV Stars of All Time by TV Guide magazine in 1996.
Early life
William Raymond Stacy Burr was born May 21, 1917, in New Westminster, British Columbia. His father William Johnston Burr (1889–1985) was a hardware salesman; his mother Minerva Annette (née Smith, 1892–1974) was a pianist and music teacher.
When Burr was six, his parents divorced. His mother moved to Vallejo, California, with him and his younger siblings Geraldine and James, while his father remained in New Westminster. Burr briefly attended San Rafael Military Academy in San Rafael, California, and graduated from Berkeley High School.
In later years, Burr freely invented stories of a happy childhood – as with many other autobiographical details he provided about his life, they are not verifiable and have no evidence to support their accuracy. In 1986, he told journalist Jane Ardmore that, when he was 12 years old, his mother sent him to New Mexico for a year to work as a ranch hand. He was already his full adult height and rather large and "had fallen in with a group of college-aged kids who didn't realize how young Raymond was, and they let him tag along with them in activities and situations far too sophisticated for him to handle". He developed a passion for growing things and joined the Civilian Conservation Corps for a year in his teens. He did acting work in his teen years, making his stage debut at age 12 with a Vancouver stock company.
Theater
Burr grew up during the Great Depression and hoped to study acting at the Pasadena Playhouse, but he was unable to afford the tuition. By his own account, which is open to question, in 1934 he joined a repertory theatre group in Toronto that toured throughout Canada, then joined another company that toured India, Australia, and England. He briefly attended Long Beach Junior College and taught for a semester at San Jose Junior College, working nights as a radio actor and singer. Resuming the verifiable part of his autobiography, Burr began his association with the Pasadena Playhouse in 1937.
Burr moved to New York in 1940 and made his first Broadway appearance in Crazy With the Heat, a two-act musical revue produced by Kurt Kasznar that quickly folded. His first starring role on the stage came in November 1942 when he was an emergency replacement in a Pasadena Playhouse production of Quiet Wedding. He became a member of the Pasadena Playhouse drama faculty for 18 months, and he performed in some 30 plays over the years. He returned to Broadway for Patrick Hamilton's The Duke in Darkness (1944), a psychological drama set during the French Wars of Religion. His performance as the loyal friend of the imprisoned protagonist led to a contract with RKO Radio Pictures.
Film
Burr appeared in more than 50 feature films between 1946 and 1957, creating an array of villains that established him as an icon of film noir. Film historian Alain Silver concluded that Burr's most significant work in the genre is in ten films: Desperate (1947), Sleep, My Love (1948), Raw Deal (1948), Pitfall (1948), Abandoned (1949), Red Light (1949), M (1951), His Kind of Woman (1951), The Blue Gardenia (1953), and Crime of Passion (1957). Silver described Burr's private detective in Pitfall as "both reprehensible and pathetic", a characterization also cited by film historian Richard Schickel as a prototype of film noir, in contrast with the appealing television characters for which Burr later became famous. "He tried to make you see the psychosis below the surface, even when the parts weren't huge," said film historian James Ursini. "He was able to bring such complexity and different levels to those characters, and create sympathy for his characters even though they were doing reprehensible things."
Other titles in Burr's film noir legacy include Walk a Crooked Mile (1948), Borderline (1950), Unmasked (1950), The Whip Hand (1951), FBI Girl (1951), Meet Danny Wilson (1952), Rear Window (1954), They Were So Young (1954), A Cry in the Night (1956), and Affair in Havana (1957). His villains were also seen in Westerns, period dramas, horror films, and adventure films.
"I was just a fat heavy", Burr told journalist James Bawden. "I split the heavy parts with Bill Conrad. We were both in our twenties playing much older men. I never got the girl but I once got the gorilla in a 3-D picture called Gorilla at Large. I menaced Claudette Colbert, Lizabeth Scott, Paulette Goddard, Anne Baxter, Barbara Stanwyck. Those girls would take one look at me and scream and can you blame them? I was drowned, beaten, stabbed and all for my art. But I knew I was horribly overweight. I lacked any kind of self esteem. At 25 I was playing the fathers of people older than me."
Burr's occasional roles on the right side of the law include the aggressive prosecutor in A Place in the Sun (1951). His courtroom performance in that film made an impression on Gail Patrick and her husband Cornwell Jackson, who had Burr in mind when they began casting the role of Los Angeles district attorney Hamilton Burger in the CBS-TV series Perry Mason.
Radio
By the age of 12, Burr was appearing in national radio dramas broadcasting in nearby San Francisco.
As a young man Burr weighed more than 300 lbs., which limited his on-screen roles. "But in radio this presented no problems, given the magnificent quality of his voice", reported The Globe and Mail. "He played romantic leads and menacing villains with equal authority, and he earned a steady and comfortable income."
Working steadily in radio since the 1940s, often uncredited, Burr was a leading player on the West Coast. He had a regular role in Jack Webb's first radio show, Pat Novak for Hire (1949), and in Dragnet (1949–50) he played Joe Friday's boss, Ed Backstrand, chief of detectives. Burr worked on other Los Angeles-based series including Suspense, Screen Directors Playhouse, Yours Truly, Johnny Dollar, Family Theater, Hallmark Playhouse and Hallmark Hall of Fame. He performed in five episodes of the experimental dramatic radio anthology series CBS Radio Workshop, and had what is arguably his best radio role in "The Silent Witness" (1957), in which his is the only voice.
In 1956 Burr was the star of CBS Radio's Fort Laramie, an adult Western drama produced, written and directed by the creators of Gunsmoke. He played the role of Lee Quince, captain of the cavalry, in the series set at a post-Civil War military post where disease, boredom, the elements and the uncharted terrain were the greatest enemies of "ordinary men who lived in extraordinary times". The half-hour transcribed program aired Sundays at 5:30 p.m. ET January 22 – October 28, 1956. Burr told columnist Sheilah Graham that he had received 1,500 fan letters after the first broadcasts, and he continued to receive letters praising the show's authenticity and presentation of human dignity.
In August 1956, CBS announced that Burr would star in the television series Perry Mason. Although the network wanted Burr to continue work on Fort Laramie as well, the TV series required an extraordinary commitment and the radio show ended.
Known for his loyalty and consciousness of history, Burr went out of his way to employ his radio colleagues in his television programs. Some 180 radio celebrities appeared on Perry Mason during the first season alone.
Television
Burr emerged as a prolific television character actor in the 1950s. He made his television debut in 1951, appearing in episodes of Stars Over Hollywood, The Bigelow Theatre, Family Theater and the debut episode of Dragnet. He went on to appear in such programs as Gruen Playhouse, Four Star Playhouse, Ford Theatre, Lux Video Theatre, Mr. and Mrs. North, Schlitz Playhouse of Stars and Playhouse 90.
Perry Mason
In 1956, Burr auditioned for the role of District Attorney Hamilton Burger in Perry Mason, a new CBS-TV courtroom drama based on the highly successful novels by Erle Stanley Gardner. Executive producer Gail Patrick Jackson had been impressed with Burr's courtroom performance in A Place in the Sun (1951), and she told Burr that he was perfect for Perry Mason but at least overweight. He went on a crash diet over the following month; he then tested as Perry Mason and was cast in the role. While Burr's test was running, Gardner reportedly stood up, pointed at the screen, and said, "That's Perry Mason." William Hopper also auditioned as Mason, but he was cast instead as private detective Paul Drake. The series also starred Barbara Hale as Della Street, Mason's secretary, William Talman as Hamilton Burger, the district attorney who loses nearly every case to Mason, and Ray Collins as homicide detective Lieutenant Arthur Tragg.
The series ran from 1957 to 1966 and made Burr a star. In the early 1960s, the show had 30 million viewers every Saturday night and Burr received 3,000 fan letters a week. Burr received three consecutive Emmy Award nominations and won the award in 1959 and 1961 for his performance as Perry Mason. The series has been rerun in syndication ever since, and was released on DVD between 2006 and 2013. Burr's character is often said never to have lost a case, although he did lose two murder cases off-screen in early episodes of the series.
Ironside
Burr moved from CBS to Universal Studios, where he played the title role in the television drama Ironside, which ran on NBC from 1967 to 1975. In the pilot episode, San Francisco Chief of Detectives Robert T. Ironside is paralyzed by a sniper during an attempt on his life and, after his recovery, uses a wheelchair for mobility, in the first crime drama show to star a policeman with a disability. The show earned Burr six Emmy nominations—one for the pilot and five for his work in the series—and two Golden Globe nominations.
Other series
After Ironside went off the air, NBC failed in two attempts to launch Burr as the star of a new series. In a two-hour television movie format, Mallory: Circumstantial Evidence aired in February 1976 with Burr again in the role of the lawyer who outwits the district attorney. Despite good reviews for Burr, the critical reception was poor, and NBC decided against developing it into a series.
In 1977, Burr starred in the short-lived TV series Kingston: Confidential as R.B. Kingston, a William Randolph Hearst-esque publishing magnate, owner of numerous newspapers and TV stations, who, in his spare time, solved crimes along with a group of employees. It was a critical failure that was scheduled opposite the extraordinarily popular Charlie's Angels. It was cancelled after 13 weeks.
Burr took on a shorter project next, playing an underworld boss in a six-hour miniseries, 79 Park Avenue.
One last attempt to launch a series followed on CBS. The two-hour premiere of The Jordan Chance aroused little interest.
On January 20, 1987, Burr hosted the television special that later served as the pilot for the long-running series Unsolved Mysteries.
Television films
In 1985, Burr was approached by producers Dean Hargrove and Fred Silverman to star in a made-for-TV movie, Perry Mason Returns. The same week, Burr recalled, he was asked to reprise the role he played in Godzilla, King of the Monsters! (1956), in a low-budget film that would be titled Godzilla 1985.
"When they asked me to do it a second time, I said, 'Certainly,' and everybody thought I was out of my mind," Burr told Tom Shales of The Washington Post. "But it wasn't the large sum of money. It was the fact that, first of all, I kind of liked 'Godzilla,' and where do you get the opportunity to play yourself 30 years later? So I said yes to both of them." Although Burr is best remembered for his role as Perry Mason, a devoted following continues to appreciate him as the actor that brought the Godzilla series to America.
He agreed to do the Mason movie if Barbara Hale returned to reprise her role as Della Street. Hale agreed, and when Perry Mason Returns aired in December 1985, her character became the defendant. The rest of the principal cast had died, but Hale's real-life son William Katt played the role of Paul Drake, Jr. The movie was so successful that Burr made a total of 26 Perry Mason television films before his death. Many were filmed in and around Denver, Colorado.
By 1993, when Burr signed with NBC for another season of Mason films, he was using a wheelchair full-time because of his failing health. In his final Perry Mason movie, The Case of the Killer Kiss, he was shown either sitting or standing while leaning on a table, but only once standing unsupported for a few seconds. Twelve more Mason movies were scheduled before Burr's death, including one scheduled to film the month he died.
As he had with the Perry Mason TV movies, Burr decided to do an Ironside reunion movie. The Return of Ironside aired in May 1993, reuniting the entire original cast of the 1967–75 series. Like many of the Mason movies, it was set and filmed in Denver.
Personal life
Physical characteristics
Burr said that he weighed 12.75 pounds (5.8 kg) at birth, and was chubby throughout his childhood. "When you're a little fat boy in public school, or any kind of school, you're just persecuted something awful", he said. His weight was always an issue for him in getting roles, and it became a public relations problem when Johnny Carson began making jokes about him during his Tonight Show monologues. Burr refused to appear as Carson's guest from then on, and told Us Weekly years later: "I have been asked a number of times to do his show and I won't do it. Because I like NBC. He's doing an NBC show. If I went on I'd have some things to say, not just about the bad jokes he's done about me, but bad jokes he does about everybody who can't fight back because they aren't there. And that wouldn't be good for NBC."
Family life
Burr married actress Isabella Ward (1919–2004) on January 10, 1948. They met in 1943 while she was a student at the Pasadena Playhouse where Burr was teaching. They met again in 1947 when she was in California with a theater company. They were married shortly before Burr began work on the 1948 film noir Pitfall. In May 1948, they appeared on stage together in a Pasadena Playhouse production based on the life of Paul Gauguin. They lived in the basement apartment of a large house in Hollywood that Burr shared with his mother and grandparents. The marriage ended within months, and Ward returned to her native Delaware. They divorced in 1952, and neither remarried.
In 1960, Burr met Robert Benevides, an actor and Korean War veteran, on the set of Perry Mason. Benevides gave up acting in 1963, and he became a production consultant for 21 of the Perry Mason TV movies. They owned and operated an orchid business and then a vineyard in California's Dry Creek Valley. They were domestic partners until Burr's death in 1993. Burr bequeathed his entire estate to Benevides, and Benevides renamed the Dry Creek property Raymond Burr Vineyards (reportedly against Burr's wishes) and managed it as a commercial enterprise. In 2017, the property was sold.
Although Burr had not revealed his homosexuality during his lifetime, it was reported in the press upon his death.
Biographical contradictions
At various times in his career, Burr and his managers and publicists offered spurious or unverifiable biographical details to the press and public. Burr's obituary in The New York Times states that he entered the US Navy in 1944, after The Duke in Darkness, and left in 1946, weighing almost . Although Burr may have served in the Coast Guard, reports of his service in the US Navy are false, as apparently are his statements that he sustained battle injuries at Okinawa.
Other false biographical details include years of college education at a variety of institutions, being widowed twice, a son who died young, world travel and success in high school athletics. Most of these claims were apparently accepted as fact by the press during Burr's lifetime, up until his death and by his first biographer, Ona Hill.
Burr reportedly was married at the beginning of World War II to an actress named Annette Sutherland—killed, Burr said, in the same 1943 plane crash that claimed the life of actor Leslie Howard. However, multiple sources have reported that no one by that name appears on any of the published passenger manifests from the flight. A son supposedly born during this marriage, Michael Evan, was said to have died of leukemia in 1953 at the age of ten. Another marriage purportedly took place in the early 1950s to a Laura Andrina Morgan—who died of cancer, Burr said, in 1955. Yet no evidence exists of either marriage, nor of a son's birth, other than Burr's own claims. As late as 1991, Burr stood by the account of this son's life and death. He told Parade that when he realized Michael was dying, he took him on a one-year tour of the United States. "Before my boy left, before his time was gone", he said, "I wanted him to see the beauty of his country and its people." After Burr's death, his publicist confirmed that Burr worked steadily in Hollywood throughout 1952, the year that he was supposedly touring the country with his son.
In the late 1950s, Burr was rumored to be romantically involved with Natalie Wood. Wood's agent sent her on public dates so she could be noticed by directors and producers, and so the men she dated could present themselves in public as heterosexuals. The dates helped to disguise Wood's relationship with Robert Wagner, whom she later married. Burr reportedly resented Warner Bros.' decision to promote her attachment to another gay actor, Tab Hunter, rather than him. Robert Benevides later said, "He was a little bitter about it. He was really in love with her, I guess."
Later accounts of Burr's life say that he hid his homosexuality to protect his career. "That was a time in Hollywood history when homosexuality was not countenanced", Associated Press reporter Bob Thomas recalled in a 2000 episode of Biography. "Ray was not a romantic star by any means, but he was a very popular figure ... If it was revealed at that time in Hollywood history it would have been very difficult for him to continue."
Arthur Marks, a producer of Perry Mason, recalled Burr's talk of wives and children: "I know he was just putting on a show. ... That was my gut feeling. I think the wives and the loving women, the Natalie Wood thing, were a bit of a cover." Dean Hargrove, executive producer of the Perry Mason TV films, said in 2006, "I had always assumed that Raymond was gay, because he had a relationship with Robert Benevides for a very long time. Whether or not he had relationships with women, I had no idea. I did know that I had trouble keeping track of whether he was married or not in these stories. Raymond had the ability to mythologize himself, to some extent, and some of his stories about his past ... tended to grow as time went by."
Hobbies and businesses
Burr had many hobbies over the course of his life: cultivating orchids and collecting wine, art, stamps, and seashells. He was very fond of cooking. He was interested in flying, sailing, and fishing. According to A&E Biography, Burr was an avid reader with a retentive memory. He was also among the earliest importers and breeders of Portuguese water dogs in the United States.
Burr developed his interest in cultivating and hybridizing orchids into a business with Benevides. Over 20 years, their company, Sea God Nurseries, had nurseries in Fiji, Hawaii, the Azores, and California, and was responsible for adding more than 1,500 new orchids to the worldwide catalog. Burr named one of them the "Barbara Hale Orchid" after his Perry Mason costar. Burr and Benevides cultivated Cabernet Sauvignon, Chardonnay, and grapes for Port wine, as well as orchids, at Burr's farmland holdings in Sonoma County, California.
In 1965, Burr purchased Naitauba, a island in Fiji, rich in seashells. There, he and Benevides oversaw the raising of copra (coconut meat) and cattle, as well as orchids. Burr planned to retire there permanently. However, medical problems made that impossible and he sold the property in 1983.
Philanthropy
Burr was a well-known philanthropist. He gave enormous sums of money, including his salaries from the Perry Mason movies, to charity. He was also known for sharing his wealth with friends. He sponsored 26 foster children through the Foster Parents' Plan or Save The Children, many with the greatest medical needs. He gave money and some of his Perry Mason scripts to the McGeorge School of Law in Sacramento, California.
Burr was an early supporter of the Bailey-Matthews National Shell Museum in Sanibel, Florida, raising funds and chairing its first capital campaign. He donated a large collection of Fijian cowries and cones from his island in Fiji.
In 1993, Sonoma State University awarded Burr an honorary doctorate. He supported medical and educational institutions in Denver, and in 1993, the University of Colorado awarded him an honorary doctorate for his acting work. Burr also founded and financed the American Fijian Foundation that funded academic research, including efforts to develop a dictionary of the language.
Burr made repeated trips on behalf of the United Service Organizations (USO). He toured both Korea and Vietnam during wartime and once spent six months touring Korea, Japan, and the Philippines. He sometimes organized his own troupe and toured bases both in the U.S. and overseas, often small installations that the USO did not serve, like one tour of Greenland, Baffin Island, Newfoundland and Labrador. Returning from Vietnam in 1965, he made a speaking tour of the U.S. to advocate an intensified war effort. As the war became more controversial, he modified his tone, called for more attention to the sacrifice of the troops, and said, "My only position on the war is that I wish it were over." In October 1967, NBC aired Raymond Burr Visits Vietnam, a documentary of one of his visits. The reception was mixed. "The impressions he came up with are neither weighty nor particularly revealing", wrote the Chicago Tribune; the Los Angeles Times said Burr's questions were "intelligent and elicited some interesting replies".
Burr had a reputation in Hollywood as a thoughtful, generous man years before much of his more-visible philanthropic work. In 1960, Ray Collins, who portrayed Lt. Arthur Tragg on the original Perry Mason series, and who was by that time often ill and unable to remember all the lines he was supposed to speak, stated, "There is nothing but kindness from our star, Ray Burr. Part of his life is dedicated to us, and that's no bull. If there's anything the matter with any of us, he comes around before anyone else and does what he can to help. He's a great star—in the old tradition."
Illness and death
During the filming of his last Perry Mason movie in the spring of 1993, Burr fell ill. A Viacom spokesperson told the media that the illness might be related to the renal cell carcinoma (malignant kidney tumor) that had been removed from Burr that February. It was determined that the cancer had spread to his liver and was at that point inoperable. Burr threw several "goodbye parties" before his death on September 12, 1993, at his Sonoma County ranch near Healdsburg. He was 76 years old.
The day after Burr's death, American Bar Association president R. William Ide III released a statement: "Raymond Burr's portrayals of Perry Mason represented lawyers in a professional and dignified manner. ... Mr. Burr strove for such authenticity in his courtroom characterizations that we regard his passing as though we lost one of our own." The New York Times reported that Perry Mason had been named second—after F. Lee Bailey, and before Abraham Lincoln, Thurgood Marshall, Janet Reno, Ben Matlock and Hillary Clinton—in a recent National Law Journal poll that asked Americans to name the attorney, fictional or not, they most admired.
Burr was interred with his parents at Fraser Cemetery, New Westminster, British Columbia. On October 1, 1993, about 600 family members and friends paid tribute to Burr at a private memorial service at the Pasadena Playhouse.
Burr bequeathed his estate to Robert Benevides, and excluded all relatives, including a sister, nieces, and nephews. His will was challenged, without success, by the two children of his late brother, James E. Burr. Benevides's attorney said that tabloid reports of an estate worth $32 million were an overestimate.
Accolades
For his work in the TV series Perry Mason, Burr received the Emmy Award for Best Actor in a Leading Role (Continuing Character) in a Dramatic Series at the 11th Primetime Emmy Awards in 1959. Nominated again in 1960, he received his second Emmy Award for Outstanding Performance by an Actor in a Series (Lead) at the 13th Primetime Emmy Awards in 1961. Burr was named Favorite Male Performer, for Perry Mason, in TV Guide magazine's inaugural TV Guide Award readers poll in 1960. He also received the second annual award in 1961.
In 1960, Burr was awarded a star on the Hollywood Walk of Fame at 6656 Hollywood Boulevard. Burr received six Emmy nominations (1968–72) for his work in the TV series Ironside. He was nominated twice, in 1969 and 1972, for the Golden Globe Award for Best Actor – Television Series Drama. A benefactor of legal education, Burr was principal speaker at the founders' banquet of the Thomas M. Cooley Law School in Lansing, Michigan, in June 1973. The Raymond Burr Award for Excellence in Criminal Law was established in his honor.
Burr was ranked #44 on TV Guides 50 Greatest TV Stars of All Time in 1996. Completed in 1996, a circular garden at the entrance to the Bailey-Matthews National Shell Museum in Sanibel, Florida, honors Burr for his role in establishing the museum. Burr was a trustee and an early supporter who chaired the museum's first capital campaign and made direct contributions from his collection. A display about Burr as an actor, benefactor and collector opened in the museum's Great Hall of Shells in 2012.
From 2000 to 2006, the Raymond Burr Performing Arts Society leased the historic Columbia Theatre from the city of New Westminster, and renamed it the Raymond Burr Performing Arts Centre. Although the nonprofit organization hoped to raise funds to renovate and expand the venue, its contract was not renewed. The group was a failed bidder when the theater was sold in 2011.
In 2008, Canada Post issued a postage stamp in its "Canadians in Hollywood" series featuring Burr. Burr received the 2009 Canadian Legends Award and a star on Canada's Walk of Fame in Toronto. The induction ceremony was held on September 12, 2009. A 2014 article in The Atlantic that examined how Netflix categorized nearly 77,000 different personalized genres found that Burr was rated as the favorite actor by Netflix users, with the greatest number of dedicated microgenres.
Theatre credits
Film credits
Radio credits
Television credits
See also
List of celebrities who own wineries and vineyards
Notes
References
External links
Raymond Burr at the MBC Encyclopedia of Television
Complete biography of Raymond Burr
Raymond Burr at Find a Grave
1917 births
1993 deaths
20th-century Canadian male actors
Berkeley High School (Berkeley, California) alumni
Canadian emigrants to the United States
Canadian expatriate male actors in the United States
Canadian gay actors
Canadian male film actors
Canadian male radio actors
Canadian male television actors
Canadian people of World War II
Canadian philanthropists
Civilian Conservation Corps people
Columbia University alumni
Deaths from cancer in California
Deaths from kidney cancer
Male actors from British Columbia
Outstanding Performance by a Lead Actor in a Drama Series Primetime Emmy Award winners
Stanford University alumni |
null | null | Blue Mountains (New South Wales) | eng_Latn | The Blue Mountains are a mountainous region and a mountain range located in New South Wales, Australia. The region borders on Sydney's metropolitan area, its foothills starting about west of centre of the state capital, close to Penrith on the outskirts of Greater Sydney region. The public's understanding of the extent of the Blue Mountains is varied, as it forms only part of an extensive mountainous area associated with the Great Dividing Range. As defined in 1970, the Blue Mountains region is bounded by the Nepean and Hawkesbury rivers in the east, the Coxs River and Lake Burragorang to the west and south, and the Wolgan and Colo rivers to the north. Geologically, it is situated in the central parts of the Sydney Basin.
The Blue Mountains Range comprises a range of mountains, plateau escarpments extending off the Great Dividing Range about northwest of Wolgan Gap in a generally southeasterly direction for about , terminating at . For about two-thirds of its length it is traversed by the Great Western Highway and the Main Western railway line. Several established towns are situated on its heights, including Katoomba, Blackheath, , and . The range forms the watershed between Coxs River to the south and the Grose and Wolgan rivers to the north. The range contains the Explorer Range and the Bell Range.
The Blue Mountains area includes the local government area of the City of Blue Mountains.
Since the early 2010s, the region's biodiversity and infrastructure has been severely affected by massive bushfires of unprecedented size and impact.
Etymology
Following European settlement of the Sydney area, the area was named the Carmarthen and Lansdowne Hills by Arthur Phillip in 1788. The Carmarthen Hills were in the north of the region and the Lansdowne Hills were in the south. The name Blue Mountains, however, was preferred and is derived from the blue tinge the range takes on when viewed from a distance. The tinge is believed to be caused by Mie scattering which occurs when incoming light with shorter wavelengths is preferentially scattered by particles within the atmosphere imparting a blue-greyish colour to any distant objects, including mountains and clouds. Volatile terpenoids emitted in large quantities by the abundant eucalyptus trees in the Blue Mountains may cause Mie scattering and thus the blue haze for which the mountains were named.
History
Geological history
A sandstone plateau dominated by a landscape of endless eucalyptus forests, the Blue Mountains are located at the top of a 470-million-year-old quartzite. In the Permian period, a shallow sea covered the region, when rivers brought shales, siltstones and mudstones. Then during the Mesozoic period, rivers dumped vast amounts of sand, burying the other sedimentary rocks. By about 1 million years ago during the Pliocene era, the Blue Mountains began to form, carving out deep valleys and steep cliffs, with an elevation of 3,900 feet. Its main distinctive natural feature are the Three Sisters, a formation of triple sandstone peaks scoured by erosion, whose rocks are lit by the glimpse of light at the time of day.
Aboriginal inhabitants
The Blue Mountains have been inhabited for millennia by the Gundungurra people, now represented by the Gundungurra Tribal Council Aboriginal Corporation based in Katoomba, and, in the lower Blue Mountains, by the Darug people, now represented by the Darug Tribal Aboriginal Corporation.
The Gundungurra creation story of the Blue Mountains tells that Dreaming creatures Mirigan and Garangatch, half fish and half reptile, fought an epic battle which scarred the landscape into the Jamison Valley.
The Gundungurra Tribal Council is a nonprofit organisation representing the Gundungurra traditional owners, promoting heritage and culture and providing a support for Gundungurra people connecting back to Country.
Gundungurra Tribal Council Aboriginal Corporation has a registered Native Title Claim since 1995 over their traditional lands, which include the Blue Mountains and surrounding areas.
Examples of Aboriginal habitation can be found in many places. In the Red Hands Cave, a rock shelter near Glenbrook, the walls contain hand stencils from adults and children. On the southern side of Queen Elizabeth Drive, at Wentworth Falls, a rocky knoll has a large number of grinding grooves created by rubbing stone implements on the rock to shape and sharpen them. There are also carved images of animal tracks and an occupation cave. The site is known as Kings Tableland Aboriginal Place and dates back 22,000 years.
Australian Colonial History
Arthur Phillip, the first governor of New South Wales, first glimpsed the extent of the Blue Mountains from a ridge at the site of today's Oakhill College, Castle Hill. He named them the Carmarthen Hills, "some forty to sixty miles distant..." and he reckoned that the ground was "most suitable for government stock". This is the location where Gidley King in 1799 established a prison town for political prisoners from Ireland and Scotland.
The first documented use of the name Blue Mountains appears in Captain John Hunter's account of Phillip's 1789 expedition up the Hawkesbury River. Describing the events of about 5 July, Hunter wrote: "We frequently, in some of the reaches which we passed through this day, saw very near us the hills, which we suppose as seen from Port Jackson, and called by the governor the Blue Mountains." During the nineteenth century the name was commonly applied to the portion of the Great Dividing Range from about Goulburn in the south to the Hunter Valley in the north, but in time it came to be associated with a more limited area.
The native Aborigines knew two routes across the mountains: Bilpin Ridge, which is now the location of Bells Line of Road between Richmond and Bell, and the Coxs River, a tributary of the Nepean River. It could be followed upstream to the open plains of the Kanimbla Valley, the type of country that farmers prize.
British settlers initially considered that fertile lands lay beyond the mountains; while in the belief of many convicts China lay beyond. However, there was little fear that the mountains might provide a means of escape since they were considered impassable. This idea was, to some extent, convenient for local authorities. An "insurmountable" barrier would deter convicts from trying to escape in that direction.
A former convict, John Wilson, may have been the first colonist to cross the Blue Mountains. It is also believed that Mathew Everingham, 1795, may have also been partly successful based on letters he wrote at the time which came to light in the late 1980s. Wilson arrived with the First Fleet in 1788 and was freed in 1792. He settled in the bush, living with the Aborigines and even functioning as an intermediary between them and the settlers. In 1797 he returned to Sydney, claiming to have explored up to a hundred miles in all directions around Sydney, including across the mountains. His descriptions and observations were generally accurate, and it is possible that he had crossed the mountains via the southern aspect at the Coxs River corridor, guided by the Aborigines.
Governor Hunter was impressed by Wilson's skills and sent him on an expedition with John Price and others in January 1798. The party crossed the Nepean River and moved southwest towards the present site of Mittagong. There they turned west and found a route along the ridge where today the Wombeyan Caves Road is located. In the process they found a way to go west of the mountains, by going around them instead of across them. In March of the same year, Wilson and Price ventured to the Camden area, and then continued further south until they encountered Thirlmere Lakes, finally almost reaching the present site of Goulburn.
It is possible that the accomplishments of this expedition were suppressed by Hunter, who may not have wanted convicts to know that there was a relatively easy way out of Sydney. Wilson was killed by Aborigines after abducting an Aboriginal woman "for his personal use", but he had accomplished much as an explorer. He was never recognised as the first person to cross the mountains, possibly because his Coxs River journey could not be verified, while his route west of Mittagong may have been the "long way around" for a colony that had its eyes fixed on the sandstone fortress west of the Nepean.
Between 1798 and 1813, many people explored various parts of the mountains, from the Bilpin Ridge to the southern regions, today the site of the Kanangra-Boyd National Park. Still, they did not find a definite route across the mountains.
The 1813 crossing of the Blue Mountains by Gregory Blaxland, William Lawson and William Charles Wentworth is officially credited as the first successful European crossing. Blaxland set out with Lawson and Wentworth on 11 May 1813 and succeeded in crossing the mountains by 31 May. They ventured as far as to what is now Mount Blaxland, just west of Coxs River.
In November 1813, Macquarie sent the surveyor George Evans on an expedition to confirm the apparent discoveries made by Blaxland and his party. He was also told to see if there existed enough arable land to justify settlement. The issue had become more urgent because the colony was in the grips of a drought.
Evans and his party reached the Fish and Macquarie rivers, and the site of Bathurst. On 7 July 1814, construction of a road across the mountains was begun by William Cox. The work was at the behest of Governor Macquarie. 30 convict labourers and 8 guards completed the road on 14 January 1815 after 27 weeks of hard work.
Since the Blue Mountains are rich in coal and shale, mining for these resources began in Hartley Vale in 1865. J.B. North ran a shale mine in the Jamison Valley in the 19th century, and other operations were set up in several places. Locations for mining activities included the Jamison Valley, the upper Grose Valley, Newnes, Glen Davis and the Asgard Swamp area outside Mount Victoria. Shale mining failed in the long run because it was not financially viable.
Climate
The climate varies with elevation. At Katoomba, () the summer average maximum temperature is around 22 °C with a few days extending into the 30s (80s–90s °F) although it is quite common to see maximum temperatures stay in the teens when east coast troughs persist. Night-time temperatures are usually in the teens but can drop to single figures at times.
During winter, the temperature is typically around 10 to 11 °C in the daytime with −1 °C or so on clear nights and 3 to 4 °C on cloudy nights. Very occasionally it will get down to −3 °C or slightly lower but usually the coldest air drains into the valleys during calm, clear nights. However, the passing of cold fronts can significantly lower the average temperature during the night and the day. The Blue Mountains is not known for particularly cold mornings compared to other areas on the Central Tablelands, such as Oberon, Bathurst and Orange. There are two to three snowfalls per year.
Annual rainfall is about in the Upper Blue Mountains with many misty days.
Geography
The predominant natural vegetation of the higher ridges is eucalyptus forest. Heath-like vegetation is present on plateau edges above cliffs. The sheltered gorges often contain temperate rainforests. There are also many hanging swamps with button grass reeds and thick, deep black soil. Wollemia nobilis, the "Wollemi pine", a relict of earlier vegetation of Gondwana, is found in remote and isolated valleys of the Wollemi National Park.
Mountain peaks
The Blue Mountains Range contains smaller mountain ranges: the Bell Range near The Bells Line of Road and north of the Grose River; the Explorer Range, south of the Grose River extending west towards Mount Victoria; the Caley Range, Erskine Range, Mount Hay Range, Paterson Range, and the Woodford Range. The major recorded peaks are:
unnamed peak ()
Mount Piddington ()
Mount Boyce ()
Mount York ()
Mount Banks ()
Mount Wilson ()
Kings Tableland ()
Narrow Neck Plateau ()
Mount Solitary ()
Mount Hay ()
Mount Irvine ()
Geology
The Blue Mountains are a dissected plateau carved in sandstone bedrock. They are now a series of ridge lines separated by gorges up to deep. The highest point in the Blue Mountains, as it is now defined, is an unnamed point with an elevation of , located north-east of . However, the highest point in the broader region that was once considered to be the Blue Mountains is Mount Bindo, with an elevation . A large part of the Blue Mountains is incorporated into the Greater Blue Mountains Area World Heritage Site, consisting of seven national park areas and a conservation reserve.
The Blue Mountains area is a distinct physiographic section of the larger Hunter-Hawkesbury Sunkland province. This is in turn a part of the larger East Australian Cordillera physiographic division.
Bushfires
The main natural disasters to afflict the area are bush fires and severe storms. In recent years the lower mountains have been subjected to a series of bush fires which have caused great loss of property but relatively little loss of life. The upper mountains had not had a major fire for some decades until December 2002 (the Blackheath Glen Fire) and November 2006 when an extensive blaze in the Grose Valley threatened several communities including Bell and Blackheath (the Lawsons Long Alley Fire). This latest fire burned for almost a month but was extinguished, mainly due to a change in the weather, without loss of human life or property. A program of winter burning seemed to have been successful in reducing fires in the upper mountains.
In recent years, the brushfires have become far more destructive and expansive than before. The region was severely damaged in the 2013 New South Wales bushfires. However, even this was dwarfed by the 2019–20 Australian bushfire season, during which the entire mountain range was devastated at a scale never seen before. According to preliminary reports, up to 80% of the World Heritage Area has burned as of January 2020, many of these being areas that had never burned in any previously observed bushfires. These destructive blazes have been linked to accelerating climate change. There have been fears that the blazes may severely reduce the biodiversity of the area and even wipe out some of the threatened species in the area, such as the regent honeyeater.
World Heritage listing
The Greater Blue Mountains Area was unanimously listed as a World Heritage Area by UNESCO on 29 November 2000, becoming the fourth area in New South Wales to be listed. The area totals roughly , including the Blue Mountains, Kanangra-Boyd, Wollemi, Gardens of Stone, Yengo, Nattai and Thirlmere Lakes National Parks, plus the Jenolan Caves Karst Conservation Reserve.
This site was chosen to be included on the World Heritage list because:
Fauna
The Greater Blue Mountains Area is inhabited by over 400 different forms of animals. Among them are rare mammal species like spotted-tailed quoll, the koala, the yellow-bellied glider, and long-nosed potoroo. There are also some rare reptiles, like the Blue Mountain water skink. There are also some dingoes in the area, which form the top predators and hunt for grey kangaroos.
Tourist attractions
The Three Sisters near Katoomba, New South Wales is one of Australia's most-photographed landmarks.
The Edge Cinema offers visitors a panoramic view of the Blue Mountains.
The Giant Stairway walking track runs down a cliff into the Jamison Valley, near the Three Sisters, providing access to nature walks through the valley.
Jenolan Caves, a series of limestone caves south west of Katoomba.
The Katoomba Scenic Railway is an incline railway now used for tourism, and originally part of the Katoomba mining tramways constructed between 1878 and 1900. The incline railway descends through sandstone cliffs, via a rock tunnel with a maximum gradient of 52 degrees. Also, at this location is the Scenic Skyway, a glass-bottom aerial cable car that traverses an arm of the Jamison Valley, and the Scenic Cableway, the steepest aerial cable car in Australia.
The Norman Lindsay Gallery and Museum at Faulconbridge, operated by The National Trust of Australia and is located inside the former Norman Lindsay homestead, Springwood.
The Toy and Railway Museum at Leura.
Knapsack Viaduct, also a location where the bell miner bird can often be heard.
Lennox Bridge, Glenbrook, on Mitchell's Pass which may be hiked (downhill) from Blaxland to Emu Plains (8 km).
Valley Heights Locomotive Depot Heritage Museum.
The Zig Zag Railway: A steam-powered railway near Lithgow.
A popular way to visit the Blue Mountains is by coach.
The last remaining apple orchard on the Shipley Plateau.
The Prime Ministers' Corridor of Oaks, at Faulconbridge, near the Great Western Highway.
Red Hands Cave in the Blue Mountains National Park near Glenbrook is one of the best examples of Aboriginal art in the area.
Lincoln's Rock in Wentworth Falls, New South Wales offers spectacular views of the Jamison Valley. Formerly known as Flat Rock, it was named after Australian mountaineer Lincoln Hall (climber) in 2013.
Recreational activity
The Blue Mountains are a popular destination for rock climbers, mountain bikers and hikers as well as canyoning and other adventure sports. These sports are well catered for by guiding companies and equipment stores located mainly in Katoomba.
Popular climbing destinations include the Centennial Glen cliffs near Blackheath and Mount Piddington near the town of Mount Victoria. Climbing is currently banned on The Three Sisters.
Mountain biking takes place mainly on the many fire trails that branch away from the main spine of the Great Western Highway, such as Narrow Neck, Anderson's Fire Trail and others.
Likewise many of the fire trails are popular with day hikers, though many dedicated walking trails exist away from the fire roads.
Canyoning in the Blue Mountains is a popular sport and caters for various skill levels. It carries inherent dangers, yet for those with the appropriate skills or those looking to take a guided trip there are many great opportunities to experience a different view of the Blue Mountains.
There are numerous abseiling options available in the Blue Mountains including single and multipitch routes. There are some restrictions though with certain areas being closed for abseiling.
Cricket is a popular sport in the Blue Mountains, with the Blue Mountains Cattle Dogs representing the district in the Western Zone Premier League, Country Plate and Presidents Cup competitions.
Gallery
See also
Aboriginal sites of New South Wales
Blue Mountains National Park
Greater Blue Mountains Area
List of mountains in Australia
Geography of Sydney
References
External links
Official Blue Mountains National Park website
Visitor's Guide to the Blue Mountains – Visit NSW
Physiographic sections
Climbing areas of Australia
Tourist attractions in Sydney
Great Dividing Range
Regions of Sydney |
null | null | C-3PO | eng_Latn | C-3PO () or See-Threepio is a humanoid robot character in the Star Wars franchise who appears in the original trilogy, the prequel trilogy and the sequel trilogy. Built by Anakin Skywalker, was designed as a protocol droid intended to assist in etiquette, customs, and translation, boasting that he is "fluent in over six million forms of communication". Along with his astromech droid counterpart and friend , provides comic relief within the narrative structure of the films, and serves as a foil. Anthony Daniels has portrayed the character in eleven of the twelve theatrical Star Wars films released to date, with the exception of Solo: A Star Wars Story, where the character does not appear.
Despite his oblivious nature, has played a pivotal role in the galaxy's history, appearing under the service of Shmi Skywalker, the Lars homestead, Padmé Amidala, Bail Organa, Raymus Antilles, Luke Skywalker, Jabba the Hutt, and Leia Organa. In the majority of depictions, physical appearance is primarily a polished gold plating with a silver plated right leg (from knee joint to ankle), although his appearance varies throughout the films; including the absence of metal coverings in The Phantom Menace, a dull copper plating in Attack of the Clones, and a red left arm in The Force Awakens. also appears frequently in both canon and Star Wars Legends continuities of novels, comic books, and video games, and was a protagonist in the animated television series Droids.
Appearances
Skywalker saga
Original trilogy
Star Wars (1977)
In Star Wars, C-3PO is introduced to the audience when he and R2-D2 are aboard the consular ship Tantive IV when it is attacked by the Imperial Star Destroyer Devastator. When R2-D2 (Kenny Baker) attempts to leave the ship to deliver a secret message from Princess Leia (Carrie Fisher) to Obi-Wan Kenobi (Alec Guinness), C-3PO follows him into an escape pod, which lands on the planet Tatooine. There, the droids are captured by Jawas, and are taken to be sold. In the process of being sold to Owen Lars (Phil Brown), C-3PO convinces his new owner to buy R2-D2 as well (after they originally bought R5-D4 and his motivator exploded on the way back to the garage). The duo ultimately lead Lars' nephew, Luke Skywalker (Mark Hamill), to Obi-Wan, for whom R2-D2 plays Leia's message. After Imperial stormtroopers destroy the Lars homestead, C-3PO and R2-D2 go along with Luke and Obi-Wan on a mission to rescue Leia, transported by smugglers Han Solo (Harrison Ford) and Chewbacca (Peter Mayhew) on board the Millennium Falcon. When the ship is taken hostage on the Death Star, C-3PO helps R2-D2 shut down the space station's trash compactor, saving their human companions' lives, before escaping on board the Falcon. When R2-D2 is damaged during the Battle of Yavin, a grief-stricken C-3PO offers to donate his own parts to help repair his counterpart. C-3PO and a repaired R2-D2 are present at the celebration of the Death Star's destruction at the end of the film.
Throughout the film C-3PO is a foil to R2-D2's antics, even when C-3PO translates R2-D2's machine speech for the audience. C-3PO is originally the property of the captain on the Tantive IV, but seems to follow R2-D2 in a relationship akin to those between human children; C-3PO often following R2-D2 around, and R2-D2 needing C-3PO to translate for him. When R2-D2 is damaged in the Battle of Yavin, C-3PO offers to donate any mechanical parts helpful in his repair; but this transference is never confirmed.
The Empire Strikes Back (1980)
In The Empire Strikes Back, C-3PO is responsible for identifying the Empire's probe droid, alerting the Rebels to the Empire's awareness of their location on the sixth planet of the Hoth system. C-3PO escapes with Han, Chewbacca, and Leia in the Millennium Falcon, while R2-D2 joins Luke in his search for Yoda (Frank Oz). During this time C-3PO and Han are often shown as foils, with C-3PO quoting odds and Han defying them. After a chase through an asteroid field, the Falcon escapes to Cloud City at Bespin.
While exploring a room in Cloud City, C-3PO is blasted by an off-camera stormtrooper. In search of C-3PO, Chewbacca heads to the Ugnaught recycling facility where he finds the dismembered parts of the droid. When Darth Vader (portrayed by David Prowse, voiced by James Earl Jones) reveals his presence to the group that same day, Chewbacca is sent into a holding cell, but is permitted to rebuild the droid, which he does poorly. Thereafter, Chewbacca carries the partially rebuilt C-3PO on his back during Han's encasement in carbonite.
With the help of the city's administrator, Lando Calrissian (Billy Dee Williams), Princess Leia, Chewbacca, and C-3PO escape the city. Having C-3PO on Chewbacca's back proves to be beneficial, in that when Boba Fett (Jeremy Bulloch) escapes the city with a frozen Han, C-3PO notifies them of pursuing stormtroopers. While escaping Vader's flagship Executor, R2-D2 begins repairing C-3PO. Upon making their way to the Rebel's rendezvous point, C-3PO is fully repaired.
Return of the Jedi (1983)
In Return of the Jedi, Luke commands C-3PO and R2-D2 to deliver a message to Jabba the Hutt from Luke, who offers 3PO's services as a translator and R2-D2's on Jabba's flying sail barge. C-3PO's first translation for the crime lord is of the bounty hunter Boushh—Leia in disguise—claiming the bounty for Chewbacca as part of her plan to rescue Han. Later, Luke infiltrates the palace and kills Jabba's rancor in a duel; whereupon Jabba transfers his court to the sail barge with Luke, Han, and Chewbacca as food for the Sarlacc, while Leia serves as Jabba's slave. When Luke attempts escape, R2-D2 tosses him his lightsaber, with which he attacks Jabba's guards. In the midst of the battle, C-3PO is attacked by Jabba's pet Salacious Crumb, who pulls out his right photoreceptor before being driven off by R2-D2. Thereafter the two escape the sail barge and are retrieved by the protagonists.
C-3PO accompanies the strike force to the Forest Moon of Endor to disable the shield generator protecting the second Death Star. When he, Han, Luke, Chewbacca, and R2-D2 are captured by the Ewoks, C-3PO is perceived to be a god by the latter. When the human prisoners are threatened by the Ewoks, Luke uses the Force to levitate the droid above the crowd as demonstration of the supposed god's ability. Later that night, C-3PO narrates the history of the Rebel Alliance's fight against the Galactic Empire to the tribe, convincing them to help the Rebels at the Battle of Endor. After the Empire's second Death Star is destroyed, C-3PO joins the protagonists on Endor in celebrating the fall of the Empire.
Prequel trilogy
The Phantom Menace (1999)
C-3PO returns in Star Wars: Episode I – The Phantom Menace, the first chapter of the Star Wars prequel trilogy, where it was revealed that nine-year-old slave Anakin Skywalker (Jake Lloyd) built him out of spare parts. C-3PO meets his future partner, R2-D2, along with Jedi Master Qui-Gon Jinn (Liam Neeson), Queen Padmé Amidala (Natalie Portman) of Naboo, and Gungan Jar Jar Binks (Ahmed Best). C-3PO and R2-D2 help perfect Anakin's podracer for the Boonta Eve Classic race.
Shortly afterwards, C-3PO becomes part of Anakin's pit crew, where he sees Anakin defeat Sebulba. C-3PO and Anakin part ways when Qui-Gon frees the boy after winning a bet with Anakin's master, Watto. Before parting from C-3PO, Anakin assures the droid that his mother Shmi (Pernilla August) will not sell him.
Attack of the Clones (2002)
In Episode II – Attack of the Clones, set 10 years later, Shmi is kidnapped by a group of Tusken Raiders. Sensing that his mother is in danger, Anakin (now portrayed by Hayden Christensen) travels with Padmé to Tatooine, where they reunite with C-3PO, now with a metallic gray-colored body. He recognizes Anakin and Padmé and presents them to Anakin's stepfamily, Owen (Joel Edgerton) and Cliegg Lars (Jack Thompson) and Beru Whitesun (Bonnie Piesse). When Anakin returns with his mother's body, C-3PO attends her funeral.
After Anakin and Padmé's visit to Tatooine, C-3PO accompanies them to the planet Geonosis to rescue Obi-Wan Kenobi (Ewan McGregor) from Sith Lord Count Dooku (Christopher Lee). Shortly afterward, he follows R2-D2 into a droid-construction site, where his head is temporarily attached to the torso of a battle droid, while the head of the droid is placed onto C-3PO's torso. Influenced by the battle droid's programming, he reluctantly participates in the film's climactic battle scene, where he is stopped by Jedi Kit Fisto (Silas Carson). Having been restored by R2-D2, he leaves Geonosis with the other protagonists. At the end of the film, he is a witness to Padmé and Anakin's marriage on Naboo.
Revenge of the Sith (2005)
In Episode III – Revenge of the Sith, C-3PO is aware of Padmé's pregnancy and, in a deleted scene, is present during a secret meeting held with Senators Bail Organa (Jimmy Smits) and Mon Mothma (Genevieve O'Reilly). After Anakin falls to the dark side of the Force and becomes the Sith Lord Darth Vader, he takes C-3PO and R2-D2 with him as he massacres of the Jedi; C-3PO rationalizes Vader's behavior by saying he has been under a great deal of stress. C-3PO accompanies Padmé to Mustafar, and witnesses Vader using the Force to choke her into unconsciousness; whereupon C-3PO and R2-D2 take her to safety. When Obi-Wan returns to their spaceship following a duel with Vader, C-3PO pilots it to Polis Massa and witnesses Padmé give birth to the Skywalker twins, Luke and Leia, and die shortly afterward. C-3PO and R2-D2 become Organa's property, and he gives them to Captain Raymus Antilles (Rohan Nichol) to serve aboard the Tantive III, and orders C-3PO's memory erased in order to protect the Skywalker children from their father and the newly created Galactic Empire.
Sequel trilogy
The Force Awakens (2015)
Anthony Daniels reprised the role of C-3PO in The Force Awakens, the first installment of the sequel trilogy. C-3PO is seen with a red left arm, but gets a gold arm at the end of the film. He has been upgraded to seven million forms of communication. In the film, he is first seen along with Leia and a Resistance team, picking up Han, the stormtrooper Finn (John Boyega), the droid BB-8, the scavenger Rey (Daisy Ridley), and Chewbacca on the planet Takodana. During their adventures, C-3PO is reunited with R2-D2, who is reactivated after having been shut off for years.
The Last Jedi (2017)
In The Last Jedi, C-3PO mostly serves as an assistant for Leia and, while she is unconscious, Resistance pilot Poe Dameron (Oscar Isaac); he spends much of the film voicing his concern over the protagonist's chances in missions such as the evacuation of D'Qar at the start of the film and Finn and Rose Tico's (Kelly Marie Tran) mission on the Mega-class Star Dreadnought Supremacy later on. At the end of the film, C-3PO has a brief reunion with Luke, who winks at him before confronting the film's villain, First Order leader Kylo Ren (Adam Driver).
According to Daniels, "In The Last Jedi I became a table decoration, which I regretted, because Threepio was worth more than that."
The Rise of Skywalker (2019)
In The Rise of Skywalker, he accompanies Rey, Finn, Poe, Chewbacca, and BB-8 to Pasaana to meet with Luke's contact, who knows about a Sith artifact; the contact turns out to be Lando. During the journey, they discover a Sith dagger and try to translate its inscriptions as they hold a clue to the whereabouts of Palpatine (Ian McDiarmid) and the Sith Eternal's fleet, the Final Order, but it is revealed that C-3PO's programming prevents him translating such a language directly. The heroes manage to find a hacker on the planet Kijimi who can bypass the protocols preventing C-3PO from translating the message, but at the cost of wiping his memories. Reactivated after he has provided the translation, C-3PO accompanies the group until they return to the Resistance camp, at which point R2-D2 is able to restore his memory from an earlier back-up created just prior to Rey's first mission. He celebrates the Sith Eternal's defeat with the rest of the Resistance at the end. In part of the film, the character appears with red eyes.
Anthology films
C-3PO makes a brief appearance alongside R2-D2 in Rogue One (2016), accompanying Rebel Alliance troops to Scarif alongside Princess Leia.
Other media
Canon
Novels
After the events of Return of the Jedi, C-3PO continues to serve Leia Organa in several novels including Bloodline.
Comics
C-3PO appears in Marvel's Star Wars comic series and accompanies Han Solo on a raid on an Imperial outpost in the miniseries Shattered Empire. The circumstances that caused his left arm to be replaced with a red one is told in Marvel's one-shot Star Wars special, C-3PO: The Phantom Limb. He also appears in Star Wars: Poe Dameron.
Legends
Novels
In April 2014, most of the licensed Star Wars novels and comics produced since the originating 1977 film Star Wars were rebranded by Lucasfilm as Star Wars Legends and declared non-canon to the franchise.
According to the Star Wars comic strip, was activated hundreds of years before the events of the original film on the planet Affa. A Guide to the Star Wars Universe affirms Affa as the droid's planet of origin, narrowing down his activation date to 112 years before A New Hope. In The Phantom Menace (1999), it is revealed that was built by a young Anakin Skywalker. In the non-canon Star Wars Tales story "Thank the Maker", Vader remembers finding the defunct droid in Watto's junk heap and guessing it is several decades old.
In Star Wars media involving the Clone Wars, C-3PO serves as Padmé Amidala's personal protocol droid. In "The New Face of War", a story in Star Wars: Republic, Queen Jamillia appoints the droid to serve as liaison to the Jedi during that campaign following the defense of Naboo and the bio-plague of Ohma-D'un by the Separatists. He is a reluctant participant in many of Padmé's adventures, including a hazardous mission during the Battle of Ilum, in which his loyalty is essential in helping Jedi Master Yoda rescue fellow Jedi Luminara Unduli and Barriss Offee.
In Expanded Universe material set after Return of the Jedi, C-3PO assists Leia, Luke, and their allies in the Alliance (and later, the New Republic) on many missions. Due to his function as a protocol droid, he most often stays with Leia Organa-Solo on Coruscant, assisting her political duties, while R2-D2 often stays with Luke at the Jedi Academy.
In The Truce at Bakura, C-3PO translates Ssi-ruuvi, the language of the Ssi-ruuk, to aid the Alliance. In The Glove of Darth Vader, C-3PO and R2-D2 are members of the Senate Planetary Intelligence Network (SPIN), and undergo transformation by the Droid Modification Team to be disguised as Kessel droids so that they could infiltrate a gathering of Imperials to find out who is planning to take control of the Empire.
In The Courtship of Princess Leia, C-3PO is led to believe that Han Solo is ancestrally the King of Corellia during Han's competition with Prince Isolder for Leia's hand in marriage. C-3PO agrees to assist Han as a counselor droid; but is shocked when Han kidnaps Leia and takes her to Dathomir. Despite this, he continues to present Han in the best possible light and writes a song for him entitled "The Virtues of King Han Solo", which he backs with the sound of a full symphony orchestra. He later discovers that Han's ancestor was merely "a pretender to the throne", but is present at Han and Leia's wedding.
Following the events of Han and Leia's marriage, C-3PO returns with them to Tatooine in Tatooine Ghost. There he helps Han and Leia in their search for an Alderaanian moss painting storing a valuable code. He also reunites with Kitster Banai and Wald, childhood friends of Anakin Skywalker from Episode I. With their help, C-3PO and the others discover Shmi Skywalker's diary, whereof Leia remarks that the droid described reminds her of C-3PO. Due to his memory wipe in Episode III, C-3PO does not mention to Han and Leia his connection to Anakin. The code is eventually found and destroyed.
In Heir to the Empire, Lando reprograms C-3PO to sound like Leia in order to hide from Imperials on Nkllon; but the Empire, led by Grand Admiral Thrawn, sees through the deception. In Dark Force Rising, C-3PO goes to Honoghr with Leia during the Thrawn Crisis to discover what had caused the plight of the Noghri. He is forced to hide with Leia and Chewbacca when Imperial forces arrive overhead, but he later returns with Leia to Coruscant, where she gives birth to Jaina and Jacen Solo. He serves as caretaker to the twins, and later Anakin Solo as well. In The Last Command, C-3PO also accompanies Han, Lando, R2-D2, Luke, Chewbacca, and Mara Jade to Wayland on a mission to destroy the Mount Tantiss storehouse. Thanks to C-3PO and some Noghri, the party recruits aid from the local population and destroys the storehouse.
In Dark Apprentice, the Solo twins escape from Chewbacca's and C-3PO's watch during a trip to a zoo on Coruscant and arrive in the planet's sub-levels. C-3PO frantically searches for them, but they had been found by King Onibald Daykim and are reunited with their parents.
In Planet of Twilight, C-3PO and R2-D2 accompany Leia on a diplomatic mission to Nam Chorios but are unable to prevent her kidnapping and fail to stop the Death Seed plague unleashed on the crew by Seti Ashgad and Dzym. Eventually rescued by Han and Lando, the two droids convey their message for help.
In The Crystal Star, C-3PO accompanies Luke and Han to Crseih Station to investigate the possible existence of Jedi trainees; but instead find Waru, a creature from another dimension, from whom they learn of Waru's alliance with Lord Hethrir and of the kidnapping of the Solo children. Hethrir is destroyed following a confrontation with Leia, Han, and Luke while Waru returns to his dimension.
In The Black Fleet Crisis, C-3PO accompanies Lando, R2-D2, and Lando's associate Lobot to investigate the runaway alien ship Teljkon Vagabond; eventually to discover that the ship contains the last vestiges of the Quella civilization.
In The New Rebellion, C-3PO, along with R2-D2 and a young mechanic named Cole Fardreamer, is instrumental in stopping Kueller from regaining power by disabling the explosive devices he had placed in a large number of droids.
In Ambush at Corellia, during the outbreak of the First Corellian Insurrection, C-3PO and Chewbacca retrieve Jaina, Jacen, and Anakin from the burning Corona House. In Showdown at Centerpoint, Threepio accompanies Luke, Gaeriel Captison, Belindi Kalenda, and Lando Calrissian to Centerpoint Station, and alerts them to the fact that the temperature inside Hollowtown is increasing to dangerously high levels. In Vision of the Future, C-3PO participates in Talon Karrde's expedition to the Kathol Sector to find Jorj Car'das, serving as a translator.
Animated series
During the Droids series, set between Revenge of the Sith and A New Hope, C-3PO and R2-D2 are traded to various masters.
In Clone Wars, C-3PO's outer plating are upgraded to a shiny bronzium coating due to his tenure as Padmé's servant. He also disapproves of combat droids, once demanding to "have a serious talk with their programmer".
Comics
In 1994, Dark Horse Comics, in its self-titled compilation series, serialized a story titled Droids, which led into a 14-issue comic series continuing the adventures of and before the events of A New Hope. Lucasfilm did not require these to tie into the animated series of the same name. Anthony Daniels and Ryder Windham also co-wrote a one-shot comic titled The Protocol Offensive (1997), based on a story by Brian Daley and with art by Igor Kordey.
In Empire's End, C-3PO and R2-D2 are nearly destroyed after they spot an Imperial installing a homing beacon on the Millennium Falcon; but are saved by Han Solo and Chewbacca. The resurrected Palpatine and his fleet discover the Alliance's existence, leading C-3PO and the others to flee to Iziz, a city on the planet of Onderon. Palpatine eventually finds them, but is mortally wounded by Han and destroyed when Empatojayos Brand sacrifices himself to save the latter's son Anakin.
In the non-canon Star Wars Tales story Storyteller, C-3PO is found partially destroyed and abandoned, years after the events of the original trilogy, by two slave boys. He tells them stories of Luke Skywalker's adventures, and they imagine the stories played out by members of their own alien race. As he is finishing his tale, a Vindar slavedriver appears, blasts C-3PO's head off, and kills Otalp. After the Vindar leave the cave, Remoh finds a lightsaber in the remains of C-3PO's body and in so doing, finds hope for his people's freedom.
Television
The US Department of Health and Human Services sponsored an approximately 1 minute long anti smoking television public service announcement. In the PSA, C-3PO finds R2-D2 holding a cigarette, and warns R2 of the dangerous nature of smoking.
Behind the scenes
Anthony Daniels did not initially agree to be cast as C-3PO, but changed his mind after reading the character's part in the script and seeing a concept painting by Ralph McQuarrie, who based his early design largely on Walter Schulze-Mittendorff's Maschinenmensch from the Fritz Lang film Metropolis (1927). The initial costume had to be roughened so it would not reflect too much light. Daniels tried on the full costume for the first time the day before shooting began, and did not see its final appearance until being shown a Polaroid photograph of it while he wore the suit. George Lucas originally envisioned the character as "more of a con man" with an American accent, but changed his mind after Daniels' voice performance outshone that of 30 other actors that tested for the role. Mel Blanc was considered for the role, but according to Daniels, Blanc told Lucas that Daniels was better for the part.
C-3PO has been played by Daniels in 11 Star Wars films—ten live action (every episode of the "Skywalker Saga" and one Anthology film) and one animated. In The Phantom Menace, a skeletal C-3PO model was puppeteered by Michael Lynch (who was also a chief model maker for all three prequel films), with Daniels providing the voice, and Lynch was removed in post-production. Daniels attempted to puppeteer the skeletal droid for Episode II, but experienced difficulties; the puppet was replaced in favor of a fully plated C-3PO. A filmed scene featured Padmé outfitting C-3PO with his plating, but it was cut to improve the pacing. From Attack of the Clones to The Rise of Skywalker, as well as Rogue One, Daniels both wore the costume and provided the voice.
Daniels also played C-3PO in the live-action segments of Star Wars Holiday Special (1978), and voiced the character in the animated portion. Daniels appeared as C-3PO at the 50th Academy Awards in 1978 and 88th Academy Awards in 2016. He also voiced C-3PO in the radio drama adaptations of the original trilogy and five Star Wars animated series: Droids (1985–86), Clone Wars (2003–05), The Clone Wars (2008–14), Rebels (2014–18) and Resistance (2018–20). He reprised his role in The Lego Movie, in which has a cameo appearance with crew members of the Millennium Falcon, and in the Star Wars Forces of Destiny episode "Beasts of Echo Base", in which his voice can be heard on Leia's comlink. Daniels also voiced in Ralph Breaks the Internet (2018) and various animated Lego Star Wars productions. Star Wars Galaxy of Adventures features some of C-3PO's dialogue from the Skywalker Saga.
Daniels reprised his role of C-3PO for the Disneyland ride Star Tours, co-created by George Lucas in 1987. He was videotaped while recording the dialogue, with the tape serving as a reference for the Imagineers to program an audio-animatronic in the queue area. Both it and the audio-animatronic R2-D2 are the original props used in filming. Daniels also donned the costume for certain video segments played before and after the simulator ride itself, also portraying an electronically disguised Ewok announcer. Daniels also appeared in the costume for various promotional materials for Star Tours, including the electronic press kit, a Disney special on Star Tours hosted by Gil Gerard, and various opening day skits. When Star Tours opened a bilingual version at Disneyland Paris, Daniels re-recorded dialogue in French. ends up in the pilot's seat of the Starspeeder 1000 in the attraction's successor, Star Tours – The Adventures Continue, in which he is also played by Daniels.
Daniels made several appearances as C-3PO on numerous TV shows and commercials, notably on a Star Wars-themed episode of The Donny and Marie Show in 1977, Disneyland's 35th Anniversary and a Star Wars-themed episode of The Muppet Show in 1980. Along with R2-D2, he also has a role as the character on Sesame Street episodes 1364 and 1396. They also hosted several documentaries including The Making of Star Wars and Star Wars: Connections.
In 1977, a photograph appeared on a Topps Star Wars trading card in which appeared to have a prominent phallus. It was later withdrawn from circulation. In 2007, the official Star Wars website hypothesized that this was caused by a part of the suit that had fallen into place just as the photograph was taken. However, in 2019 Daniels clarified that the costume had become compromised during 's oil bath in A New Hope; the warm liquid had caused the costume to separate, leading to "an over-exposure of plastic in that region". He further speculated in his 2019 autobiography that an artist had painted in the extended appendage.
Cultural influence
C-3PO was inducted into the Robot Hall of Fame in 2004 in Pittsburgh, Pennsylvania. A lifesize replica can be seen at the Carnegie Science Center in Pittsburgh.
In 2012, the United States Navy built a robot called the Autonomous Shipboard Humanoid which they claim was modeled on . The robot was built for the purpose of extinguishing fires, but has similar movement abilities to the Star Wars droid by being able to climb ladders and carry hoses. Also similar to , it can respond to human gestures, but at a much more basic level.
One of NASA programs to collaborate with the private sector in the development of commercial space transportation it's named 'C3PO', acronym for Commercial Crew & Cargo Program Office.
Nicolas Ghesquière, creative director of Louis Vuitton, and Rodarte have created fashion designs inspired by .
The cover photo of the August 2015 issue of GQ featured comedian Amy Schumer wearing a replica of Princess Leia's bikini and sucking one of fingers. Lucasfilm and Disney objected to this image and tweeted a public statement that it was unauthorized.
In the Italian version of the original trilogy, was named "D-3BO". Anglicisms were not common in the Italian language during the 1970s and 1980s, and the names of various characters were changed to be easier to pronounce and recognize for Italian speakers. Some of these changes were reverted in the dubbing of the prequel and sequel trilogies, where the original name was used instead. Other name changes in the Italian version of the original trilogy include R2-D2, Han Solo, Leia Organa and Darth Vader, who were called respectively "C1-P8", "Ian Solo", "Leila Organa" and "Dart Fener" (or "Lord Fener").
See also
R2-D2
BB-8
Notes
References
Footnotes
Citations
External links
https://starwars.fandom.com/wiki/C-3PO
Characters created by George Lucas
Film characters introduced in 1977
Fictional linguists
Male characters in film
Male characters in television
Star Wars droid characters
Star Wars Anthology characters
Star Wars Skywalker Saga characters
Star Wars: The Clone Wars characters
Star Wars literary characters |
null | null | Rohan (Middle-earth) | eng_Latn | Rohan is a kingdom of horsemen, the Rohirrim, in J. R. R. Tolkien's fantasy setting of Middle-earth. It is mainly a grassland, and provides its ally Gondor with cavalry. The Rohirrim call their land the Mark or the Riddermark, the name recalling that of the Kingdom of Mercia, the region of Western England where Tolkien lived.
Rohan is grounded in Anglo-Saxon tradition, poetry, and linguistics, specifically in its Mercian dialect, in everything but its use of horses. Tolkien used Old English for the kingdom's language and names, pretending that this was in translation of Rohirric. Meduseld, King Théoden's hall, is modelled on Heorot, the great hall in Beowulf.
In The Lord of the Rings, Rohan is important in the action, first against the wizard Saruman in the Battle of the Hornburg, and then in the climactic Battle of the Pelennor Fields. There, Théoden leads the Rohirrim to victory against the forces of Mordor; he is killed when his horse falls, but his niece Éowyn kills the leader of the Ringwraiths.
In Tolkien's works
Etymology
Tolkien's own account, in an unsent letter, gives both the fictional and the actual etymologies of Rohan:
Geography
In Tolkien's Middle-earth, Rohan is an inland realm. Its countryside is described as a land of pastures and lush tall grassland which is frequently windswept. The meadows contain "many hidden pools, and broad acres of sedge waving above wet and treacherous bogs" that water the grasses. Karen Wynn Fonstad calculated Rohan to be 52,763 square miles (136,656 km2) in area (slightly larger than England).
Borders
Rohan is bordered to the north by the Fangorn forest, home to the Ents (tree-giants) led by Treebeard, and by the great river Anduin, called Langflood by the Rohirrim. To the northeast are the walls of Emyn Muil. After the War of the Ring, the kingdom is extended northwards over the Limlight to the borders of Lothlórien. To the east are the mouths of the River Entwash, and the Mering Stream, which separated Rohan from the Gondorian province of Anórien, known to the Rohirrim as Sunlending. To the south lie the White Mountains (Ered Nimrais). To the west are the rivers Adorn and Isen, where Rohan borders the land of the Dunlendings. To the northwest, just under the southern end of the Misty Mountains, lies the walled circle of Isengard around the ancient tower of Orthanc; at the time of the War of the Ring, it had been taken over by the evil wizard Saruman. The area of the western border where the Misty Mountains and the White Mountains drew near to each other is known as the Gap of Rohan.
Capital
The capital of Rohan is the fortified town of Edoras, on a hill in a valley of the White Mountains. "Edoras" is Old English for "enclosures". The town of Edoras was built by Rohan's second King, Brego son of Eorl the Young. The hill on which Edoras is built stands in the mouth of the valley of Harrowdale. The river Snowbourn flows past the town on its way east towards the Entwash. The town is protected by a high wall of timber.
Meduseld, the Golden Hall of the Kings of Rohan, is in the centre of the town at the top of the hill. "Meduseld", Old English for "mead hall", is meant to be a translation of an unknown Rohirric word with the same meaning. Meduseld is based on the mead hall Heorot in Beowulf; it is a large hall with a thatched roof that appears golden from far off. The walls are richly decorated with tapestries depicting the history and legends of the Rohirrim, and it serves as a house for the King and his kin, a meeting hall for the King and his advisors, and a gathering hall for ceremonies and festivities. It is at Meduseld that Aragorn, Gimli, Legolas, and Gandalf meet with King Théoden. Legolas describes Meduseld in a line that directly translates a line of Beowulf, "The light of it shines far over the land", representing líxte se léoma ofer landa fela. The hall is anachronistically described as having louvres to remove the smoke, derived from William Morris's 1889 The House of the Wolfings.
Other settlements
Upstream from Edoras, deeper into Harrowdale, are the hamlets of Upbourn and Underharrow. At the head of Dunharrow (from Old English Dûnhaerg, "the heathen fane on the hillside") is a refuge, Firienfeld, in the White Mountains. Aldburg, capital of the Eastfold, is the original settlement of Eorl the Young. The Hornburg, a major fortress guarding the western region, is in Helm's Deep, a valley in the White Mountains.
Regions
The kingdom of Rohan, also called the Mark, is primarily divided into two regions, the East-mark and the West-mark. They are each led by a marshal of the kingdom. Rohan's capital, Edoras, lies in a small but populous region in the centre south of the kingdom, the Folde. In an earlier concept, Rohan's capital region was called the King's Lands, of which the Folde was a sub-region to the south-east of Edoras. North of the Folde, the boundary between the East-mark and West-mark runs along the Snowbourn River and the Entwash. Most of the rest of Rohan's population is spread along the foothills of the White Mountains in both directions from the Folde. In the West-mark the Westfold extends along the mountains to Helm's Deep (the defensive centre of Westfold) and to the Gap of Rohan. Beyond the Gap of Rohan lies the West Marches, the kingdom's far west borderland. The Eastfold extends along the White Mountains in the opposite direction (and was thus a part of the East-mark). It is bound by the Entwash to the north. Its eastern borderland is called the Fenmarch; beyond this lies the Kingdom of Gondor.
The centre of Rohan is a large plain, divided by the Entwash into the East Emnet and the West Emnet. These regions fell respectively into the East-mark and the West-mark. The northernmost region of Rohan, and the least populous, is the Wold. The Field of Celebrant (named for a synonym of the River Silverlode), even further north, is added to Rohan after the War of the Ring.
Culture
People
The Dúnedain of Gondor and the Rohirrim are distantly related, having descended from the same place. Unlike the inhabitants of Gondor, who are portrayed as enlightened and highly civilized, the Rohirrim are shown as being at a lower level of enlightenment.
The names and many details of Rohirric culture are derived from Germanic cultures, particularly that of the Anglo-Saxons and their Old English language, towards which Tolkien felt a strong affinity. Anglo-Saxon England was defeated by the cavalry of the Normans at the Battle of Hastings, and some Tolkien scholars have speculated that the Rohirrim are Tolkien's wishful version of an Anglo-Saxon society that retained a "rider culture", and would have been able to resist such an invasion. The Tolkien scholar Tom Shippey notes that Tolkien derived the emblem of the House of Éorl, a "white horse upon green", from the Uffington White Horse carved into the grass of the chalk downs in England.
While Tolkien represents the Rohirrim with Anglo-Saxon culture and language, their ancestors are given Gothic attributes. The names of Rhovanion's royal family, (the ancestors of the Rohirrim), include such names as Vidugavia, Vidumavi and Vinitharya, which are of Gothic origin. Vidugavia specifically has been seen as an synonym for Vitiges, king of the Ostrogoths in Italy from 536 to 540. Tolkien saw this as a parallel with the real-world relationship between Old English and Gothic.
In response to a query about clothing styles in Middle-earth, Tolkien wrote:
Horses and warfare
The armies of Rohan were largely horsemen. The basic tactical unit was the éored, Old English for "a unit of cavalry, a troop", which at the time of the War of the Ring had a nominal strength of 120 riders.
In time of war, every able man was obliged to join the Muster of Rohan. Rohan was bound by the Oath of Éorl to help Gondor in times of peril, and the latter asked for their aid through the giving of the Red Arrow. This has a historical antecedent in the Old English poem Elene, in which Constantine the Great summoned an army of mounted Visigoths to his aid against the Huns by sending an arrow as a "token of war". Gondor could also call the Rohirrim in need by lighting the warning beacons of Gondor, seven (in the Lord of the Ring movie thirteen) signal fires along the White Mountains from Minas Tirith to the Rohan border: Amon Dîn, Eilenach, Nardol, Erelas, Min-Rimmon, Calenhad and Halifirien.
At the start of the War of the Ring a Full Muster would have been over 12,000 riders. Among the horses of the Rohirrim were the famed mearas, the noblest and fastest horses that ever roamed Arda. It was because of the close affiliation with horses, both in war and peace, that they received their name. Rohirrim (or more properly Rochirrim) is Sindarin for "Horse-lords," and Rohan (or Rochand) means "Land of the Horse-lords".
Language
Tolkien generally called the language simply "the language of Rohan" or "of the Rohirrim". The adjectival form "Rohirric" is common; Tolkien once also used "Rohanese". Like many languages of Men, it is akin to Adûnaic, the language of Númenóreans, and therefore to the Westron or Common Speech.
The Rohirrim called their homeland the Riddermark, a modernization by Tolkien of Old English Riddena-mearc, meaning, according to the Index to The Lord of the Rings, "the border country of the knights"; also Éo-marc, the Horse-mark, or simply the Mark. They call themselves the Éorlingas, the Sons of Éorl. Tolkien rendered the language of the Riders of Rohan, Rohirric, as the Mercian dialect of Old English. Even words and phrases that were printed in modern English showed a strong Old English influence. This solution occurred to Tolkien when he was searching for an explanation of the Eddaic names of the dwarves already published in The Hobbit. Tolkien, a philologist, with a special interest in Germanic languages, pretended that the names and phrases of Old English were translated from Rohirric, just as the English used in The Shire was supposedly translated from Middle-earth's Westron or Common Speech. Examples include éored and mearas. The Riders' names for the cunningly-built tower of Isengard, Orthanc, and for the Ents, the tree-giants of Fangorn forest, are similarly Old English, both being found in the phrase orþanc enta geweorc, "cunning work of giants" in the poem The Ruin, though Shippey suggests that Tolkien may have chosen to read the phrase also as "Orthanc, the Ent's fortress".
In The Two Towers, chapter 6, the Riders of Rohan are introduced before they are seen, by Aragorn, who chants in the language of the Rohirrim words "in a slow tongue unknown to the Elf and the Dwarf", a lai that Legolas senses "is laden with the sadness of Mortal Men". The song is called the Lament of the Rohirrim. To achieve a resonant sense of the lost past, the now-legendary time of a peaceful alliance of the Horse-lords with the realm of Gondor, Tolkien adapted the short Ubi sunt ("Where are they?") passage of the Old English poem The Wanderer.
"Thus spoke a forgotten poet long ago in Rohan, recalling how tall and fair was Eorl the Young, who rode down out of the North," Aragorn explains, after singing the Lament.
History
Early history
In the 13th century of the Third Age, the Kings of Gondor made close alliances with the Northmen of Rhovanion, a people said in The Lord of the Rings to be akin to the Three Houses of Men (later the Dúnedain) from the First Age. In the 21st century, a remnant tribe of such Northmen, the Éothéod, moved from the valleys of Anduin to the northwest of Mirkwood, disputing with the Dwarves over the treasure-hoard of Scatha the dragon.
In 2509, Cirion the Steward of Gondor summoned the Éothéod to help repel an invasion of Men from Rhûn and Orcs from Mordor. Eorl the Young, lord of the Éothéod, answered the summons, arriving unexpected at a decisive battle on the Field of Celebrant, routing the orc army. As a reward, Éorl was given the Gondorian province of Calenardhon (except for Isengard).
Kingdom of Rohan
Eorl the Young founded the Kingdom of Rohan in the former Calenardhon; the royal family was known as the House of Eorl. The first line of kings lasted for 249 years, until the ninth king Helm Hammerhand died. His sons had been killed earlier, and his nephew Fréaláf Hildeson began the second line of kings, which lasted until the end of the Third Age. The two lines of kings are buried in two lines of grave mounds below the royal hall at Edoras, like those at Gamla Uppsala in Sweden, or Sutton Hoo in England.
In 2758, Rohan was invaded by Dunlendings under Wulf, son of Freca, of mixed Dunland and Rohan blood. The King, Helm Hammerhand, took refuge in the Hornburg until help from Gondor and Dunharrow arrived a year later. Soon after this Saruman took over Isengard, and was welcomed as an ally.
War of the Ring
Saruman used his influence through the traitor Grima Wormtongue to weaken Théoden. Saruman then launched an invasion of Rohan, with victory in early battles at the Fords of Isen, killing Théoden's son, Théodred. Saruman was defeated at the Battle of the Hornburg, where the tree-like Huorns came from the forest of Fangorn to help the Rohirrim.
Théoden then rode with his army to Minas Tirith, helping to break its siege in the Battle of the Pelennor Fields and killing the leader of the Haradrim, but was killed when his horse fell. He was succeeded by his nephew Éomer. His niece Éowyn and the hobbit Merry Brandybuck killed the Lord of the Nazgûl.
Éomer rode with the armies of Gondor to the Black Gate of Mordor and took part in the Battle of the Morannon against the forces of Sauron. At this time, the destruction of the Ruling Ring in Mount Doom ended the battle and the war. Éowyn married Faramir, Prince of Ithilien.
Analysis
The critic Jane Chance writes that Théoden is transformed by Gandalf into a good bold "Germanic king"; she contrasts this with the failure of "the proud Beorhtnoth" in the Old English poem The Battle of Maldon. In her view, in the account of the battle of Helm's Deep, the fortress of the Riddermark, Tolkien is emphasising the Rohirrim's physical prowess.
The critic Tom Shippey notes that the Riders of Rohan are, despite Tolkien's protestations, much like the ancient English (the Anglo-Saxons), but that they differed from the ancient English in having a culture based on horses. They use many Old English words related to horses; their name for themselves is Éotheod, horse-people, and the names of riders like Éomund, Éomer, and Éowyn begin with the word for "horse", eo[h]. In Shippey's view, a defining virtue of the Riders is panache, which he explains means both "the white horsetail on [Éomer's] helm floating in his speed" and "the virtue of sudden onset, the dash that sweeps away resistance." Shippey notes that this allows Tolkien to display Rohan both as English, based on their Old English names and words like éored ("troop of cavalry"), and as "alien, to offer a glimpse of the way land shapes people".
Shippey states further that "the Mark" (or the Riddermark), the land of the Riders of Rohan – all of whom have names in the Mercian dialect of Old English, was once the usual term for central England, and it would have been pronounced and written "marc" rather than the West Saxon "mearc" or the Latinized "Mercia".
The Tolkien scholar Thomas Honegger, agreeing with Shippey's description of the Rohirrim as "Anglo-Saxons on Horseback", calls the sources for them "quite obvious to anyone familiar with Anglo-Saxon literature and culture". The resemblances, according to Honegger, include masterly horsemanship, embodying the Old English saying Éorl sceal on éos boge, éored sceal getrume rídan ("The leader shall on horse's back, warband shall ride in a body"). The Riders are a Germanic warrior-society, exemplifying the "northern heroic spirit", like the Anglo-Saxons. But the "crucial" fact is the language; Honneger notes that Tolkien had represented Westron speech as modern English; since Rohan spoke a related but older language, Old English was the natural choice in the same style; Tolkien's 1942 table of correspondences also showed that the language of the people of Dale was represented by Norse. Honegger notes that this does not equate the Rohirrim with the Anglo-Saxons (on horseback or not), but it does show a strong connection, making them "the people most dear to Tolkien and all medievalists."
Jane Ciabattari writes on BBC Culture that Lady Éowyn's fear of being caged rather than "doing great deeds" by riding to battle with the Rohirrim resonated with 1960s feminists, contributing to the success of Lord of the Rings at that time.
Portrayal in adaptations
For Peter Jackson's The Lord of the Rings film trilogy, the Poolburn Reservoir in Central Otago, New Zealand was used for Rohan scenes. The theme for Rohan is played on a Hardanger fiddle. A fully realised set for Edoras was built on Mount Sunday in the upper reaches of the Rangitata Valley, near Erewhon in New Zealand. Some of the set was built digitally, but the main buildings atop the city were built on location; the mountain ranges in the background were part of the actual location shot. The interiors of buildings such as the Golden Hall, however, were located on soundstages in other parts of New Zealand; when the camera is inside of the Golden Hall, looking out the open gates, the image of the on-set Edoras set is digitally inserted into the door-frame. The site was known among the cast and crew for being extremely windy, as can be seen during the film and DVD interviews. After filming, Mount Sunday was returned to its original state.
Notes
References
Primary
This list identifies each item's location in Tolkien's writings.
Secondary
Works cited
reprinted in Hammond, Wayne G.; Scull, Christina (2005), The Lord of the Rings: A Reader's Companion, HarperCollins, .
External links
Fictional elements introduced in 1954
Middle-earth realms
Fictional kingdoms |
null | null | Static random-access memory | eng_Latn | Static random-access memory (static RAM or SRAM) is a type of random-access memory (RAM) that uses latching circuitry (flip-flop) to store each bit. SRAM is volatile memory; data is lost when power is removed.
The term 'static' differentiates SRAM from DRAM (dynamic random-access memory) which must be periodically refreshed. SRAM is faster and more expensive than DRAM; it is typically used for the cache and internal registers of a CPU while DRAM is used for a computer's main memory.
History
Semiconductor bipolar SRAM was invented in 1963 by Robert Norman at Fairchild Semiconductor. MOS SRAM was invented in 1964 by John Schmidt at Fairchild Semiconductor. It was a 64-bit MOS p-channel SRAM.
The SRAM was the main driver behind any new CMOS-based technology fabrication process since 1959 when CMOS was invented. In 1965, Arnold Farber and Eugene Schlig, working for IBM, created a hard-wired memory cell, using a transistor gate and tunnel diode latch. They replaced the latch with two transistors and two resistors, a configuration that became known as the Farber-Schlig cell. In 1965, Benjamin Agusta and his team at IBM created a 16-bit silicon memory chip based on the Farber-Schlig cell, with 80 transistors, 64 resistors, and 4 diodes.
Applications and uses
Characteristics
Though it can be characterized as volatile memory SRAM exhibits data remanence.
SRAM offers a simple data access model and does not require a refresh circuit. Performance and reliability are good and power consumption is low when idle.
Since SRAM requires more transistors per bit to implement, it is less dense and more expensive than DRAM and also has a higher power consumption during read or write access. The power consumption of SRAM varies widely depending on how frequently it is accessed.
Embedded use
Many categories of industrial and scientific subsystems, automotive electronics, and similar embedded systems, contain SRAM which, in this context, may be referred to as ESRAM. Some amount (kilobytes or less) is also embedded in practically all modern appliances, toys, etc. that implement an electronic user interface.
SRAM in its dual-ported form is sometimes used for real-time digital signal processing circuits.
In computers
SRAM is also used in personal computers, workstations, routers and peripheral equipment: CPU register files, internal CPU caches and external burst mode SRAM caches, hard disk buffers, router buffers, etc. LCD screens and printers also normally employ SRAM to hold the image displayed (or to be printed). SRAM was used for the main memory of most early personal computers such as the ZX80, TRS-80 Model 100 and Commodore VIC-20.
Hobbyists
Hobbyists, specifically home-built processor enthusiasts, often prefer SRAM due to the ease of interfacing. It is much easier to work with than DRAM as there are no refresh cycles and the address and data buses are often directly accessible. In addition to buses and power connections, SRAM usually requires only three controls: Chip Enable (CE), Write Enable (WE) and Output Enable (OE). In synchronous SRAM, Clock (CLK) is also included.
Types of SRAM
Non-volatile SRAM
Non-volatile SRAM (nvSRAM) has standard SRAM functionality, but they save the data when the power supply is lost, ensuring preservation of critical information. nvSRAMs are used in a wide range of situationsnetworking, aerospace, and medical, among many otherswhere the preservation of data is critical and where batteries are impractical.
Pseudo SRAM
Pseudostatic RAM (PSRAM) has a DRAM storage core, combined with a self refresh circuit. They appear externally as a slower SRAM. They have a density/cost advantage over true SRAM, without the access complexity of DRAM.
Cypress Semiconductor's HyperRAM is a type of PSRAM supporting a JEDEC compliant 8 pin HyperBusor Octal xSPI interface. These low-power devices have high performance and low pin-count, making them ideal for applications requiring external RAM as buffer memory.
By transistor type
Bipolar junction transistor (used in TTL and ECL) very fast but with high power consumption
MOSFET (used in CMOS) low power and very common today
By flip-flop type
Binary SRAM
Ternary SRAM
By function
Asynchronous independent of clock frequency; data in and data out are controlled by address transition. Examples include the ubiquitous 28-pin 8K × 8 and 32K × 8 chips (often but not always named something along the lines of 6264 and 62C256 respectively), as well as similar products up to 16 Mbit per chip.
Synchronous all timings are initiated by the clock edges. Address, data in and other control signals are associated with the clock signals.
In 1990s, asynchronous SRAM used to be employed for fast access time. Asynchronous SRAM was used as main memory for small cache-less embedded processors used in everything from industrial electronics and measurement systems to hard disks and networking equipment, among many other applications. Nowadays, synchronous SRAM (e.g. DDR SRAM) is rather employed similarly like Synchronous DRAMDDR SDRAM memory is rather used than asynchronous DRAM. Synchronous memory interface is much faster as access time can be significantly reduced by employing pipeline architecture. Furthermore, as DRAM is much cheaper than SRAM, SRAM is often replaced by DRAM, especially in the case when large volume of data is required. SRAM memory is however much faster for random (not block / burst) access. Therefore, SRAM memory is mainly used for CPU cache, small on-chip memory, FIFOs or other small buffers.
By feature
Zero bus turnaround (ZBT) the turnaround is the number of clock cycles it takes to change access to the SRAM from write to read and vice versa. The turnaround for ZBT SRAMs or the latency between read and write cycle is zero.
syncBurst (syncBurst SRAM or synchronous-burst SRAM) features synchronous burst write access to the SRAM to increase write operation to the SRAM
DDR SRAM Synchronous, single read/write port, double data rate I/O
Quad Data Rate SRAMSynchronous, separate read and write ports, quadruple data rate I/O
Integrated on chip
SRAM may be integrated as RAM or cache memory in micro-controllers (usually from around 32 bytes up to 128 kilobytes), as the primary caches in powerful microprocessors, such as the x86 family, and many others (from 8 KB, up to many megabytes), to store the registers and parts of the state-machines used in some microprocessors (see register file), on application-specific ICs, or ASICs (usually in the order of kilobytes) and in Field Programmable Gate Array and Complex Programmable Logic Device
Design
A typical SRAM cell is made up of six MOSFETs, and is often called a SRAM cell. Each bit in the cell is stored on four transistors (M1, M2, M3, M4) that form two cross-coupled inverters. This storage cell has two stable states which are used to denote 0 and 1. Two additional access transistors serve to control the access to a storage cell during read and write operations. In addition to 6T SRAM, other kinds of SRAM chips use 4, 8, 10 (4T, 8T, 10T SRAM), or more transistors per bit. Four-transistor SRAM is quite common in stand-alone SRAM devices (as opposed to SRAM used for CPU caches), implemented in special processes with an extra layer of polysilicon, allowing for very high-resistance pull-up resistors. The principal drawback of using 4T SRAM is increased static power due to the constant current flow through one of the pull-down transistors (M1 or M2).
This is sometimes used to implement more than one (read and/or write) port, which may be useful in certain types of video memory and register files implemented with multi-ported SRAM circuitry.
Generally, the fewer transistors needed per cell, the smaller each cell can be. Since the cost of processing a silicon wafer is relatively fixed, using smaller cells and so packing more bits on one wafer reduces the cost per bit of memory.
Memory cells that use fewer than four transistors are possible but, such 3T or 1T cells are DRAM, not SRAM (even the so-called 1T-SRAM).
Access to the cell is enabled by the word line (WL in figure) which controls the two access transistors M5 and M6 which, in turn, control whether the cell should be connected to the bit lines: BL and BL. They are used to transfer data for both read and write operations. Although it is not strictly necessary to have two bit lines, both the signal and its inverse are typically provided in order to improve noise margins.
During read accesses, the bit lines are actively driven high and low by the inverters in the SRAM cell. This improves SRAM bandwidth compared to DRAMs in a DRAM, the bit line is connected to storage capacitors and charge sharing causes the bit line to swing upwards or downwards. The symmetric structure of SRAMs also allows for differential signaling, which makes small voltage swings more easily detectable. Another difference with DRAM that contributes to making SRAM faster is that commercial chips accept all address bits at a time. By comparison, commodity DRAMs have the address multiplexed in two halves, i.e. higher bits followed by lower bits, over the same package pins in order to keep their size and cost down.
The size of an SRAM with m address lines and n data lines is 2m words, or 2m × n bits. The most common word size is 8 bits, meaning that a single byte can be read or written to each of 2m different words within the SRAM chip. Several common SRAM chips have 11 address lines (thus a capacity of 211 = 2,048 = 2k words) and an 8-bit word, so they are referred to as "2k × 8 SRAM".
The dimensions of an SRAM cell on an IC is determined by the minimum feature size of the process used to make the IC.
SRAM operation
An SRAM cell has three different states: standby (the circuit is idle), reading (the data has been requested) or writing (updating the contents). SRAM operating in read mode and write modes should have "readability" and "write stability", respectively. The three different states work as follows:
Standby
If the word line is not asserted, the access transistors M5 and M6 disconnect the cell from the bit lines. The two cross-coupled inverters formed by M1M4 will continue to reinforce each other as long as they are connected to the supply.
Reading
In theory, reading only requires asserting the word line WL and reading the SRAM cell state by a single access transistor and bit line, e.g. M6, BL. However, bit lines are relatively long and have large parasitic capacitance. To speed up reading, a more complex process is used in practice: The read cycle is started by precharging both bit lines BL and BL, to high (logic 1) voltage. Then asserting the word line WL enables both the access transistors M5 and M6, which causes one bit line BL voltage to slightly drop. Then the BL and BL lines will have a small voltage difference between them. A sense amplifier will sense which line has the higher voltage and thus determine whether there was 1 or 0 stored. The higher the sensitivity of the sense amplifier, the faster the read operation. As the NMOS is more powerful, the pull-down is easier. Therefore, bit lines are traditionally precharged to high voltage. Many researchers are also trying to precharge at a slightly low voltage to reduce the power consumption.
Writing
The write cycle begins by applying the value to be written to the bit lines. If we wish to write a 0, we would apply a 0 to the bit lines, i.e. setting BL to 1 and BL to 0. This is similar to applying a reset pulse to an SR-latch, which causes the flip flop to change state. A 1 is written by inverting the values of the bit lines. WL is then asserted and the value that is to be stored is latched in. This works because the bit line input-drivers are designed to be much stronger than the relatively weak transistors in the cell itself so they can easily override the previous state of the cross-coupled inverters. In practice, access NMOS transistors M5 and M6 have to be stronger than either bottom NMOS (M1, M3) or top PMOS (M2, M4) transistors. This is easily obtained as PMOS transistors are much weaker than NMOS when same sized. Consequently, when one transistor pair (e.g. M3 and M4) is only slightly overridden by the write process, the opposite transistors pair (M1 and M2) gate voltage is also changed. This means that the M1 and M2 transistors can be easier overridden, and so on. Thus, cross-coupled inverters magnify the writing process.
Bus behavior
RAM with an access time of 70 ns will output valid data within 70 ns from the time that the address lines are valid. Some SRAM have a "page mode" where words of a page (256, 512, or 1024 words) can be read sequentially with a significantly shorter access time (typically approximately 30 ns). The page is selected by setting the upper address lines and then words are sequentially read by stepping through the lower address lines.
Production challenges
With the introduction of the FinFET transistor implementation of SRAM cells, they started to suffer from increasing inefficiencies in cell sizes. Over the last 30 years (from 1987 to 2017) with a steadily decreasing transistor size (node size) the footprint-shrinking of the SRAM cell itself slowed down, making it harder to pack cells more densely.
Besides issues with size a significant challenge of modern SRAM cells is a static current leakage. The current, that flows from positive supply, through the cell, and to the ground, increases exponentially when the cell's temperature rises. The cell power drain occurs in both active and idle states wasting useful energy without any useful work done. Even though in the last 20 years the issue was partially addressed by Data Retention Voltage technique (DRV) with reduction rates ranging from 5 to 10, the decrease in node size caused reduction rates to fell to about 2.
With these two issues it became more challenging to develop energy-efficient and dense SRAM memories prompting the semiconductor industry to look for alternatives such as STT-MRAM and F-RAM.
Research
In 2019 a French institute reported on a research of an IoT-purposed 28nm fabricated IC. It was based on fully depleted silicon on insulator-transistors (FD-SOI), had two-ported SRAM memory rail for synchronous/asynchronous accesses, and selective virtual ground (SVGND). The study has claimed reaching an ultra-low SVGND current in a "sleep" and read modes by finely tuning its voltage.
See also
Flash memory
Miniature Card, a discontinued SRAM memory card standard
In-memory processing
References
Computer memory
Types of RAM |
null | null | Sphynx cat | eng_Latn | The Sphynx cat (Sfyngks "/ˈsfɪŋks/") is a breed of cat known for its lack of fur. Hairlessness in cats is a naturally occurring genetic mutation, and the Sphynx was developed through selective breeding of these animals, starting in the 1960s.
According to breed standards, the skin should have the texture of chamois leather, as it has fine hairs, or the cat may be completely hairless. Whiskers may be present, either whole or broken, or may be totally absent. The cats have a narrow, long head and webbed feet. Their skin is the color that their fur would be, and all the usual cat markings (solid, point, van, tabby, tortie, etc.) may be found on the Sphynx cat's skin. Because they have no fur, Sphynx cats lose body heat more readily than coated cats, making them both warm to the touch and prone to seeking out warm places.
Breed standards
The breed standards are defined by The International Cat Association (TICA):
Wedge-shaped heads with prominent cheekbones
Large, lemon-shaped eyes
Very large ears with hair on inside, but soft down on outside base
Well-muscled, powerful neck of medium length
Medium length torso, barrel-chested, and full, round abdomen, sometimes called a pot belly
Paw pads thicker than other cats, giving the appearance of walking on cushions
Whiplike, tapering tail from body to tip, (sometimes with fur all over tail or a puff of fur on the tip, like a lion)
Muscular body
Behavior
Sphynx are known for their extroverted behavior. They display a high level of energy, intelligence, curiosity and affection for their owners. They are one of the more dog-like breeds of cats, frequently greeting their owners at the door and friendly when meeting strangers.
History of the cat breed
The contemporary breed of Sphynx cat, also known as the Canadian Sphynx, is distinct from the Russian hairless cat breeds, like Peterbald and Donskoy.
Although hairless cats have been reported throughout history, breeders in Europe have been developing the Sphynx breed since the early 1960s. Two different sets of hairless felines discovered in North America in the 1970s provided the foundation cats for what was shaped into the existing Sphynx breed.
The current American and European Sphynx breed is descended from two lines of natural mutations:
Dermis and Epidermis (1975) barn cats from the Pearson family of Wadena, Minnesota
Bambi, Punkie and Paloma (1978) stray cats found in Toronto, Ontario, Canada, and raised by Shirley Smith
Toronto
The Canadian Sphynx breed was started in 1966 in Toronto, Ontario when a hairless kitten named Prune was born to a black and white domestic shorthair queen (Elizabeth). The kitten was mated with its mother (something called backcrossing), which produced one more naked kitten. Together with a few naked kittens found later, the cat Prune was the first attempt to create a hairless breed.
After purchasing these cats in 1966 and initially referring to them as "Moonstones" and "Canadian Hairless", Ridyadh Bawa, a science graduate of the University of Toronto, combined efforts with his mother Yania, a longtime Siamese breeder, and Keese and Rita Tenhoves to develop a breed of cats which was subsequently renamed as Sphynx. The Bawas and the Tenhoves were the first individuals able to determine the autosomal recessive nature of the Sphynx gene for hairlessness while also being successful in transforming this knowledge into a successful breeding program with kittens which were eventually capable of reproducing. The Tenhoves were initially able to obtain for the new breed provisional showing status through the Cat Fanciers' Association (CFA) but ultimately had the status revoked in 1971, when it was felt by the CFA Board that the breed had concerns over fertility.
The first breeders had rather vague ideas about Sphynx genetics and faced a number of problems. The genetic pool was very limited and many kittens died. There was also a problem with many of the females suffering convulsions. In 1978, cat breeder Shirley Smith found three hairless kittens on the streets of her neighborhood. In 1983, she sent two of them to Dr. Hugo Hernandez in the Netherlands to breed the two kittens, named Punkie and Paloma, to a white Devon Rex named Curare van Jetrophin. The resulting litter produced five kittens: two males from this litter (Q. Ramses and Q. Ra) were used, along with Punkie's half-sister, Paloma.
Minnesota
The first noted naturally occurring foundation Sphynx originated as hairless stray barn cats in Wadena, Minnesota, at the farm of Milt and Ethelyn Pearson. The Pearsons identified hairless kittens occurring in several litters of their domestic shorthair barn cats in the mid-1970s. Two hairless female kittens born in 1975 and 1976, Epidermis and Dermis, were sold to Oregon breeder Kim Mueske, and became an important part of the Sphynx breeding program. Also working with the Pearson line of cats was breeder Georgiana Gattenby of Brainerd, Minnesota, who outcrossed with Cornish Rex cats.
Genetics and breeding
Other hairless breeds may have body shapes or temperaments that differ from those of Sphynx standards. There are, for example, new hairless breeds, including the Don Sphynx and the Peterbald from Russia, which arose from their own spontaneous gene mutations. The standard for the Sphynx differs between cat associations such as The International Cat Association (TICA), Fédération Internationale Féline (FIFE) and Cat Fanciers' Association (CFA).
Breeding
In 2010, DNA analysis confirmed that Sphynx hairlessness was produced by a different allele of the same gene that produces the short curly hair of the Devon Rex (termed the "re" allele), with the Sphynx's allele being incompletely dominant over the Devon allele and both being recessive to the wild type. The Sphynx's allele is termed "hr", for hairless. The only allowable outcross breeds in the CFA are now the American Shorthair and Domestic Shorthair. Other associations may vary, and the Russian Blue is a permitted outcross in the Governing Council of the Cat Fancy (GCCF). In Europe, mainly the Devon Rex has been used for outcrosses.
Genetics
The Sphynx's hairlessness is produced by a mutation in the same gene that produces the short curly coat of the Devon Rex. Moreover, it was found that the curly coat of Selkirk Rex cats is also associated with this gene.
The gene encodes keratin 71 (KRT71) and is responsible for the keratinization of the hair follicle. The Sphynx's mutation leads to a complete loss of function where the structure of the hair is damaged so that the hair can be formed but is easily dislodged. In the Devon Rex mutation, a residual activity of the protein still exists. The Selkirk Rex allele (sadr) is dominant over the wild type gene, which is dominant over the Devon Rex allele (re) and the Sphynx (hr), which forms an allelic series of : KRT71SADRE > KRT71+ > KRT71re > KRT71hr.
Care
Although Sphynx cats lack a coat to shed or groom, they are not maintenance-free. Body oils, which would normally be absorbed by the hair, tend to build up on the skin. As a result, regular bathing is usually necessary weekly or bi-weekly. Care should be taken to limit the Sphynx cat's exposure to outdoor sunlight at length, as they can develop sunburn and skin damage similar to that of humans. In general, Sphynx cats should never be allowed outdoors unattended, as they have limited means to conserve body heat when it is cold. In some climates, owners provide coats or other clothing in the winter to help them conserve body heat.
While they lack much of the fur of other cat breeds, Sphynxes are not necessarily hypoallergenic. Allergies to cats are triggered by a protein called Fel d1, not cat hair itself. Fel d1 is a protein primarily found in cat saliva and sebaceous glands. Those with cat allergies may react worse to direct contact with Sphynx cats than other breeds. Even though reports exist that some people with allergies successfully tolerate Sphynx cats, they are fewer than those who have allergic reactions, according to David Rosenstreich, MD, the director of the Division of Allergy and Immunology at the Albert Einstein College of Medicine and Montefiore Medical Center in the Bronx, New York City, New York. These positive reports may be cases of desensitizing, wherein the "hairless" cat gave the owner optimism to try to own a cat, eventually leading to the positive situation of their own adaptation.
Sphynx cats can also have more ear wax than most hairy domestic cats because they have little to no hair in their ears. Dirt, skin oils (sebum) and ear wax accumulates in the ears, and needs to be cleaned out on a weekly basis, usually before bath time.
The Sphynx breed also tends to accumulate oils and debris under the nails as well as the skin fold above the nail due to the lack of fur, so, like the ears, the nails and surrounding skin folds need to be cleaned properly as well. Due to these factors the breed may require more grooming than a typical domestic cat with fur. Specialty products for this cat have been developed, though they still require more grooming than most breeds.
Health issues
The Canadian Sphynx faces challenges because of its lack of protective fur. Skin cancer may be a problem if exposed to sunlight for long durations of time.
The lack of hair can cause health issues with kittens in the first weeks of life because of susceptibility to respiratory infections. Reputable breeders should not let their kittens go to new homes without being at least 14 weeks of age to ensure the kitten is mature enough to cope in a new environment.
Hypertrophic cardiomyopathy
The breed does have instances of the genetic disorder hypertrophic cardiomyopathy (HCM). Other domestic cat breeds prone to HCM include Persian, Ragdoll, Norwegian Forest cat, Siberian cats, British Shorthair and Maine Coon; however, any domestic cat including mixed breeds can acquire HCM. Studies are being undertaken to understand the links in breeding and the disorder. Cats are screened for HCM disease with echocardiography (ultrasound of the heart), as well as with additional tests determined by the veterinarian cardiologist including electrocardiogram (EKG, ECG), chest radiographs (X-rays), and/or blood tests.
The Sphynx cat has a high rate of heart disease, either as HCM or mitral valve dysplasia. In a 2012 study of 114 Sphynx cats, 34% were found to have an abnormal heart, with 16 cats having mitral valve dysplasia and 23 cats having HCM. These prevalences were found in cats with an average age of 2.62 years. Males cats developed more severe disease than female cats and often developed it earlier, at an average age of 19 months for males and 29 months for females. Since the prevalence of genetic heart disease is high in this breed, many breeders will recommend screening for HCM yearly.
As HCM progresses into an advanced stage, cats may experience congestive heart failure (CHF) or thromboembolism.
Congenital myasthenic syndrome
Congenital myasthenic syndrome (CMS) previously referred to as muscular dystrophy, myopathy or spasticity, is a type of inherited neuromuscular disorder associated with alpha-dystroglycan deficiency, found in Sphynx and in Devon Rex cats as well as variants of these breeds, which can occur between the first 3 to 23 weeks of their life. This condition has also been described, but is rarely seen. Cats affected by CMS show generalized muscle weakness and fatigue, as well as ventroflexion of the head and neck, head bobbing, and scapulae protrusion.
References
Cat breeds
Cat breeds originating in Canada
Hairless cat breeds
Rare cat breeds
Culture of Ontario
Articles containing video clips |
null | null | New Spain | eng_Latn | New Spain, officially the Viceroyalty of New Spain (, ), or Kingdom of New Spain, was an integral territorial entity of the Spanish Empire, established by Habsburg Spain during the Spanish colonization of the Americas. Its jurisdiction comprised a huge area that included what are now Mexico, much of the Southwestern U.S. and California in North America, Central America, northern parts of South America, and several Pacific Ocean archipelagos.
After the 1521 Spanish conquest of the Aztec empire, conqueror Hernán Cortés named the territory New Spain, and established the new capital of Mexico City on the site of the Tenochtitlan, the capital of the Mexica (Aztec) Empire. Central Mexico became the base of expeditions of exploration and conquest, expanding the territory claimed by the Spanish Empire. With the political and economic importance of the conquest, the crown asserted direct control over the densely populated realm. The crown established New Spain as a viceroyalty in 1535, appointing as viceroy Antonio de Mendoza, an aristocrat loyal to the monarch rather than the conqueror Cortés. New Spain was the first of the viceroyalties that Spain created, the second being Peru in 1542, following the Spanish conquest of the Inca Empire. Both New Spain and Peru had dense indigenous populations at conquest as a source of labor and material wealth in the form of vast silver deposits, discovered and exploited beginning in the mid 1500s.
New Spain developed highly regional divisions based on local climate, topography, distance from the capital and the Gulf Coast port of Veracruz, size and complexity of indigenous populations, and the presence or absence of mineral resources. Central and southern Mexico had dense indigenous populations, each with complex social, political, and economic organization, but no large-scale deposits of silver to draw Spanish settlers. By contrast, the northern area of Mexico was arid and mountainous, a region of nomadic and semi-nomadic indigenous populations, which do not easily support human settlement. In the 1540s, the discovery of silver in Zacatecas attracted Spanish mining entrepreneurs and workers, to exploit the mines, as well as crown officials to ensure the crown received its share of revenue. Silver mining became integral not only to the development of New Spain, but to the Spanish crown, which depended on the revenues from silver mining, vastly enriching Spain and transformed the global economy. New Spain's port of Acapulco became the New World terminus of the transpacific trade with Asia via the Manila galleon. The New Spain became a vital link between Spain's New World empire and its East Indies empire.
From the beginning of the 19th century, the kingdom fell into crisis, aggravated by the 1808 Napoleonic invasion of Iberia and the forced abdication of the Bourbon monarch, Charles IV. This resulted in the political crisis in New Spain and much of the Spanish Empire in 1808, which ended with the government of Viceroy José de Iturrigaray. Later, it gave rise to the Conspiracy of Valladolid and the Conspiracy of Querétaro. This last one was the direct antecedent of the Mexican War of Independence. At its conclusion in 1821, the viceroyalty was dissolved and the Mexican Empire was established. Former royalist military officer turned insurgent for independence Agustín de Iturbide would be crowned as emperor.
The Crown and the Viceroyalty of New Spain
The Kingdom of New Spain was established on August 18, 1521, following the Spanish conquest of the Aztec Empire, as a New World kingdom dependent on the Crown of Castile. The initial funds for exploration came from Queen Isabella. Although New Spain was a dependency of Castile, it was a kingdom and not a colony, subject to the presiding monarch on the Iberian Peninsula.
The monarch had sweeping power in the overseas territories, with not just sovereignty over the realm but also property rights. All power over the state came from the monarch, and via the Patronato real, the crown had sweeping powers over the church as well. The Viceroyalty of New Spain was created by royal decree on October 12, 1535, in the Kingdom of New Spain with a viceroy appointed as the king's "deputy" or substitute. This was the first New World viceroyalty and one of only two that the Spanish empire administered in the continent until the 18th-century Bourbon Reforms.
Territorial extent of the overseas Spanish Empire
The Spanish Empire comprised the territories in the north overseas 'Septentrion', from North America and the Caribbean, to the Philippine, Mariana and Caroline Islands. At its greatest extent, the Spanish crown claimed on the mainland of the Americas much of North America south of Canada, that is: all of present-day Mexico and Central America except Panama; most of present-day United States west of the Mississippi River, plus the Floridas. The Spanish West Indies, settled prior to the conquest of Mexico, also came under New Spain's jurisdiction: (Cuba, Hispaniola (comprising the modern states of Haiti and the Dominican Republic), Puerto Rico, Jamaica, the Cayman Islands, Trinidad, and the Bay Islands). New Spain's jurisdiction also included the Spanish East Indies in Asia and Oceania, (the Philippine Islands, the Mariana Islands, the Caroline Islands, parts of Taiwan, and parts of the Moluccas). Although asserting sovereignty over this vast realm, it did not effectively control large swaths. Other European powers, including England, France, and the Netherlands established colonies in territories Spain claimed.
Much of what was called in the U.S. the "Spanish borderlands", is territory that did not attract many Spanish settlers, with less dense indigenous populations and apparently lacking in mineral wealth. The huge deposits of gold in California were discovered immediately after it was incorporated into the U.S. following the Mexican-American War (1846–48). The territory in the colonial was considered more marginal to Spanish interests than the most densely populated and lucrative areas of central Mexico. To shore up its claims in North America in the eighteenth century as other power encroached on its claims, the crown sent expeditions to the Pacific Northwest, which explored and claimed the coast of what is now British Columbia and Alaska. Religious missions and fortified presidios were established to shore up Spanish control on the ground. On the mainland, the administrative units included Las Californias, that is, the Baja California peninsula, still part of Mexico and divided into Baja California and Baja California Sur; Alta California (present-day Arizona, California, Nevada, Utah, western Colorado, and southern Wyoming); (from the 1760s) Louisiana (including the western Mississippi River basin and the Missouri River basin); Nueva Extremadura (the present-day states of Coahuila and Texas); and Santa Fe de Nuevo México (parts of Texas and New Mexico).
Political organization
Viceroyalty
The Viceroyalty was administered by a viceroy residing in Mexico City and appointed by the Spanish monarch, who had administrative oversight of all of these regions, although most matters were handled by the local governmental bodies, which ruled the various regions of the viceroyalty. First among these were the audiencias, which were primarily superior tribunals, but which also had administrative and legislative functions. Each of these was responsible to the Viceroy of New Spain in administrative matters (though not in judicial ones), but they also answered directly to the Council of the Indies.
Captaincies general and governorates
The Captaincy Generals were the second-level administrative divisions and these were relatively autonomous from the viceroyalty.
Santo Domingo (1535); Philippines (1574); Puerto Rico (1580); Cuba (1608); Guatemala (1609); Yucatán (1617);Commandancy General of the Provincias Internas (1776) (analogous to a dependent captaincy general). Two governorates, third-level administrative divisions, were established, the Governorate of Spanish Florida, (Spanish: La Florida) and the Governorate of Spanish Louisiana (Spanish: Luisiana).
High courts
The high courts, or audiencias, were established in major areas of Spanish settlement. In New Spain the high court was established in 1527, prior to the establishment of the viceroyalty. The First Audiencia was headed by Hernán Cortés's rival Nuño de Guzmán, who used the court to deprive Cortés of power and property. The crown dissolved the First Audiencia and established the Second Audiencia.
The audiencias of New Spain were Santo Domingo (1511, effective 1526, predated the Viceroyalty); Mexico (1527, predated the Viceroyalty); Panama (1st one, 1538–1543); Guatemala (1543); Guadalajara (1548); Manila (1583). Audiencia districts further incorporated the older, smaller divisions known as governorates (gobernaciones, roughly equivalent to provinces), which had been originally established by conquistador-governors known as adelantados. Provinces which were under military threat were grouped into captaincies general, such as the Captaincies General of the Philippines (established 1574) and Guatemala (established in 1609) mentioned above, which were joint military and political commands with a certain level of autonomy. (The viceroy was captain-general of those provinces that remained directly under his command).
Local-level administration
At the local level there were over two hundred districts, in both indigenous and Spanish areas, which were headed by either a corregidor (also known as an alcalde mayor) or a cabildo (town council), both of which had judicial and administrative powers. In the late 18th century the Bourbon dynasty began phasing out the corregidores and introduced intendants, whose broad fiscal powers cut into the authority of the viceroys, governors and cabildos. Despite their late creation, these intendancies so affected the formation of regional identity that they became the basis for the nations of Central America and the first Mexican states after independence.
Intendancies 1780s
As part of the sweeping eighteenth-century administrative and economic changes known as the Bourbon Reforms, the Spanish crown created new administrative units called intendancies, to strengthen central control over the viceroyalty. Some measures aimed to break the power of local elites in order to improve the economy of the empire. Reforms included the improvement of the public participation in communal affairs, distribution of undeveloped lands to the indigenous and Spaniards, end the corrupt practices of local crown officials, encourage trade and mining, and establish a system of territorial division similar to the model created by the government of France, already adopted in Spain. The establishment of intendancies found strong resistance by the viceroyalties and general captaincies similar to the opposition in the Iberian Peninsula when the reform was adopted. Royal audiencias and ecclesiastical hierarchs opposed the reform for its interventions in economic issues, by its centralist politics, and the forced ceding of many of their functions to the intendants. In New Spain, these units generally corresponded to the regions or provinces that had developed earlier in the center, South, and North. Many of the intendancy boundaries became Mexican state boundaries after independence. The intendancies were created between 1764 and 1789, with the greatest number in the mainland in 1786: 1764 Havana (later subdivided); 1766 New Orleans; 1784 Puerto Rico; 1786 Mexico, Veracruz, Puebla de Los Angeles, Guadalajara, Guanajuato, Zacatecas, San Luis Potosí, Sonora, Durango, Oaxaca, Guatemala, San Salvador, Comayagua, León, Santiago de Cuba, Puerto Príncipe; 1789 Mérida.
History of New Spain
The history of mainland New Spain spans three hundred years from the Spanish conquest of the Aztec Empire (1519-1521) to the collapse of Spanish rule in the Mexican War of Independence (1810-1821).
Conquest era (1521–1535)
The Caribbean islands and early Spanish explorations around the circum-Caribbean region had not been of major political, strategic, or financial importance until the conquest of the Aztec Empire in 1521. However, important precedents of exploration, conquest, and settlement and crown rule had been initially worked out in the Caribbean, which long affected subsequent regions, including Mexico and Peru. The indigenous societies of Mesoamerica brought under Spanish control were of unprecedented complexity and wealth compared to what the conquerors had encountered in the Caribbean. This presented both an important opportunity and a potential threat to the power of the Crown of Castile, since the conquerors were acting independent of effective crown control. The societies could provide the conquistadors, especially Hernán Cortés, a base from which the conquerors could become autonomous, or even independent, of the Crown. Cortés had already defied orders that curtailed his ambition of an expedition of conquest. He was spectacularly successful in gaining indigenous allies against the Aztec Empire, with the indispensable aid of indigenous cultural translator, Marina, also known as Malinche, toppling the rulers of the Aztec empire. Cortés then divvied up the spoils of war without crown authorization, including grants of labor and tribute of groups of indigenous, to participants in the conquest.
As a result, the Holy Roman Emperor and King of Spain, Charles V created the Council of the Indies in 1524 as the crown entity to oversee the crown's interests in the New World. Since the time of the Catholic Monarchs, central Iberia was governed through councils appointed by the monarch with particular jurisdictions. The creation of the Council of the Indies became another, but extremely important, advisory body to the monarch.
The crown had already created the Casa de Contratación (House of Trade) in 1503 to regulate contacts between Spain and its overseas possessions. A key function was to gather information about navigation to make trips less risky and more efficient. Philip II sought systematic information about his overseas empire and mandated reports, known as the Relaciones geográficas, describing topography, economic conditions, and populations, among other information. They were accompanied by maps of the area discussed, many of which were drawn by indigenous artists. The Francisco Hernández Expedition (1570–77), the first scientific expedition to the New World, was sent to gather information on medicinal plants and practices.
The crown created the first mainland high court, or Audiencia, in 1527 to regain control of the administration of New Spain from Cortés, who as the premier conqueror of the Aztec empire, was ruling in the name of the king but without crown oversight or control. An earlier Audiencia had been established in Santo Domingo in 1526 to deal with the Caribbean settlements. That court, housed in the Casa Reales in Santo Domingo, was charged with encouraging further exploration and settlements with the authority granted it by the crown. Management by the Audiencia, which was expected to make executive decisions as a body, proved unwieldy. In 1535, Charles V of Spain appointed Don Antonio de Mendoza as the first Viceroy of New Spain, an aristocrat loyal to the crown, rather than the conqueror Hernán Cortés, who had embarked on the expedition of conquest and distributed spoils of the conquest without crown approval. Cortés was instead awarded a vast, entailed estate and a noble title.
Christian evangelization
Spanish conquerors saw it as their right and their duty to convert indigenous populations to Christianity. Because Christianity had played such an important role in the Reconquista (Christian reconquest) of the Iberian peninsula from the Muslims, the Catholic Church in essence became another arm of the Spanish government, since the crown was granted sweeping powers over ecclesiastical affairs in its overseas territories. The Spanish Crown granted it a large role in the administration of the state, and this practice became even more pronounced in the New World, where prelates often assumed the role of government officials. In addition to the Church's explicit political role, the Catholic faith became a central part of Spanish identity after the conquest of last Muslim kingdom in the peninsula, the Emirate of Granada, and the expulsion of all Jews who did not convert to Christianity.
The conquistadors brought with them many missionaries to promulgate the Catholic religion. Amerindians were taught the Roman Catholic religion and the Spanish language. Initially, the missionaries hoped to create a large body of Amerindian priests, but were not successful. They did work to keep the Amerindian cultural aspects that did not violate the Catholic traditions, and a syncretic religion developed. Most Spanish priests committed themselves to learn the most important Amerindian languages (especially during the 16th century) and wrote grammars so that the missionaries could learn the languages and preach in them. This was similar to practices of French colonial missionaries in North America.
At first, conversion of indigenous peoples seemed to happen rapidly. The missionaries soon found that most of the natives had simply adopted "the god of the heavens," as they called the Christian god, as another one of their many gods. While they often held the Christian god to be an important deity because it was the god of the victorious conquerors, they did not see the need to abandon their old beliefs. As a result, a second wave of missionaries began an effort to completely erase the old beliefs, which they associated with the ritualized human sacrifice found in many of the native religions. They eventually prohibited this practice, which had been common before Spanish colonization. In the process many artifacts of pre-Columbian Mesoamerican culture were destroyed. Hundreds of thousands of native codices were burned, native priests and teachers were persecuted, and the temples and statues of the old gods were torn down. The missionaries even prohibited some foods associated with the native religions, such as amaranth.
Many clerics, such as Bartolomé de las Casas, also tried to protect the natives from de facto and actual enslavement to the settlers, and obtained from the Crown decrees and promises to protect native Mesoamericans, most notably the New Laws. But the royal government was too far away to fully enforce these, and settler abuses against the natives continued, even among the clergy. Eventually, the Crown declared the natives to be legal minors and placed under the guardianship of the Crown, which was responsible for their indoctrination. It was this status that barred the native population from the priesthood. During the following centuries, under Spanish rule, a new culture developed that combined the customs and traditions of the indigenous peoples with that of Catholic Spain. The Spaniards had numerous churches and other buildings constructed in the Spanish style by native labor, and named their cities after various saints or religious topics, such as San Luis Potosí (after Saint Louis) and Vera Cruz (the True Cross).
The Spanish Inquisition, and its New Spanish counterpart, the Mexican Inquisition, continued to operate in the viceroyalty until Mexico declared its independence in 1821. This resulted in the execution of more than 30 people during the colonial period. During the 17th and 18th centuries, the Inquisition worked with the viceregal government to block the diffusion of liberal ideas during the Enlightenment, as well as the revolutionary republican and democratic ideas of the United States War of Independence and the French Revolution.
Sixteenth-century founding of Spanish cities
During the first twenty years after the conquest, before the establishment of the viceroyalty, some of the important cities of the colonial era that remain important today were founded. Even before the 1535 establishment of the viceroyalty of New Spain, conquerors in central Mexico founded new Spanish cities and embarked on further conquests, a pattern that had been established in the Caribbean. In central Mexico, they transformed the Aztec capital of Tenochtitlan into the main settlement of the territory; thus, the history of Mexico City is of huge importance to the whole colonial enterprise. Spaniards founded new settlements in Puebla de los Angeles (founded 1531) at the midway point between the Mexico City (founded 1521–24) and the Caribbean port of Veracruz (1519). Colima (1524), Antequera (1526, now Oaxaca City), and Guadalajara (1532) were all new Spanish settlements. North of Mexico City, the city of Querétaro was founded (ca. 1531) in a region known as the Bajío, a major zone of commercial agriculture. Guadalajara was founded northwest of Mexico City (1531–42) and became the dominant Spanish settlement in the region. West of Mexico City the settlement of Valladolid (Michoacan) was founded (1529–41). In the densely populated indigenous South, as noted, Antequera (1526) became the center of Spanish settlement in Oaxaca; Santiago de Guatemala was founded in 1524; and in Yucatán, Mérida (1542) was founded inland, with Campeche founded in 1541 as a small, Caribbean port. Sea trade flourished between Campeche and Veracruz. The discovery of silver in Zacatecas in the far north was a transformative event in the history of New Spain and the Spanish Empire, with silver becoming the primary driver of the economy. The city of Zacatecas was founded in 1547 and Guanajuato, the other major mining region, was founded in 1548, deep in the territory of the nomadic and fierce Chichimeca, whose resistance to Spanish presence became known as the protracted conflict of the Chichimeca War. The silver was so valuable to the crown that waging a fifty-year war was worth doing. Other Spanish cities founded before 1600 were the Pacific coast port of Acapulco (1563), Durango (1563), Saltillo (1577), San Luis Potosí (1592), and Monterrey (1596). The cities were outposts of European settlement and crown control, while the countryside was almost exclusively inhabited by the indigenous populations.
Later mainland expansion
During the 16th century, many Spanish cities were established in North and Central America. Spain attempted to establish missions in what is now the southern United States, including Georgia and South Carolina between 1568 and 1587. These efforts were mainly successful in the region of present-day Florida, where the city of St. Augustine was founded in 1565. It is the oldest European city in the United States.
Upon his arrival, Viceroy Don Antonio de Mendoza vigorously took to the duties entrusted to him by the King and encouraged the exploration of Spain's new mainland territories. He commissioned the expeditions of Francisco Vásquez de Coronado into the present day American Southwest in 1540–1542. The Viceroy commissioned Juan Rodríguez Cabrillo in the first Spanish exploration up the Pacific Ocean in 1542–1543. Cabrillo sailed far up the coast, becoming the first European to see present-day California, now part of the United States. The Viceroy also sent Ruy López de Villalobos to the Spanish East Indies in 1542–1543. As these new territories became controlled, they were brought under the purview of the Viceroy of New Spain. Spanish settlers expanded to Nuevo Mexico, and the major settlement of Santa Fe was founded in 1610.
The establishment of religious missions and military presidios on the northern frontier became the nucleus of Spanish settlement and the founding of Spanish towns.
Pacific expansion and the Philippine trade
Seeking to develop trade between the East Indies and the Americas across the Pacific Ocean, Miguel López de Legazpi established the first Spanish settlement in the Philippine Islands in 1565, which became the town of San Miguel (present-day Cebu City). Andrés de Urdaneta discovered an efficient sailing route from the Philippine Islands to Mexico which took advantage of the Kuroshio Current. In 1571, the city of Manila became the capital of the Spanish East Indies, with trade soon beginning via the Manila-Acapulco Galleons. The Manila-Acapulco trade route shipped products such as silk, spices, silver, porcelain and gold to the Americas from Asia. The first census in the Philippines was founded in 1591, based on tributes collected. The tributes count the total founding population of Spanish-Philippines as 667,612 people, of which: 20,000 were Chinese migrant traders, at different times: around 16,500 individuals were Latino soldier-colonists who were cumulatively sent from Peru and Mexico and they were shipped to the Philippines annually, 3,000 were Japanese residents, and 600 were pure Spaniards from Europe, there was also a large but unknown number of Indian Filipinos, the rest of the population were Malays and Negritos. Thus, with merely 667,612 people, during this era, the Philippines was among the most sparsely populated lands in Asia.
Despite the sparsity of the Philippine population, it was profitable for Mexico City which used it as a transhipment point of cheap Asian products like Silk and Porcelain, however, due to the larger quantity of products from Asia it became a point of contention with the mercantilist policies of mainland Spain which supported manufacturing based on the capital instead of the colonies, in which case the Manila-Mexico commercial alliance was at odds against Madrid. The importance of the Philippines to the Spanish empire can be seen by its creation as a separate Captaincy-General. Products brought from Asia were sent to Acapulco then overland to Veracruz, and then shipped to Spain aboard the West Indies Fleets. Later they were traded across Europe. Several cities and towns in the Philippines were founded as Presidios commanded by Spanish officers and staffed by Mexican and Peruvian soldiers who were mostly forcefully conscripted vagrants, estranged teenagers, petty criminals, rebels or political exiles at Mexico and Peru and were thus a rebellious element among the Spanish colonial apparatus in the Philippines.
Since the Philippines was at the center of a crescent from Japan to Indonesia, it alternated into periods of extreme wealth congregating to the location, to periods where it was the arena of constant warfare waged between it and the surrounding nation(s). This left only the fittest and strongest to survive and serve out their military service. There was thus high desertion and death rates which also applied to the native Filipino warriors and laborers levied by Spain, to fight in battles all across the archipelago and elsewhere or build galleons and public works. The repeated wars, lack of wages, dislocation, and near starvation were so intense, that almost half of the soldiers sent from Latin America and the warriors and laborers recruited locally either died or disbanded to the lawless countryside to live as vagabonds among the rebellious natives, escaped enslaved Indians (from India) and Negrito nomads, where they race-mixed through rape or prostitution, which increased the number of Filipinos of Spanish or Latin American descent, but were not the children of valid marriages. This further blurred the racial caste system Spain tried so hard to maintain in the towns and cities. These circumstances contributed to the increasing difficulty of governing the Philippines. Due to these, the Royal Fiscal of Manila wrote a letter to King Charles III of Spain, in which he advises to abandon the colony, but this was successfully opposed by the religious and missionary orders that argued that the Philippines was a launching pad for further conversions in the Far East. Due to the missionary nature of the Philippine colony, unlike in Mexico where most immigrants were of a civilian nature, most settlers in the Philippines were either soldiers, merchants or clergy and were overwhelmingly male.
At times, non-profitable war-torn Philippine colony survived on an annual subsidy paid by the Spanish Crown and often procured from taxes and profits accumulated by the Viceroyalty of New Spain (Mexico), mainly paid by annually sending 75 tons of precious Silver Bullion, gathered from and mined from Potosi, Bolivia where hundreds of thousands of Incan lives were regularly lost while being enslaved to the Mit'a system. Unfortunately, the silver mined through the cost of many lives and being a precious metal barely made it to the starving or dying Spanish, Mexican, Peruvian and Filipino soldiers who were stationed in presidios across the archipelago, struggling against constant invasions, while it was sought after by Chinese, Indian, Arab and Malay merchants in Manila who traded with the Latinos for their precious metal in exchange for silk, spices, pearls and aromatics. Trade and immigration was not just aimed towards the Philippines, however. It also went the opposite direction, to the Americas, from rebellious Filipinos, especially the exiled Filipino royalties, who were denied their traditional rights by new Spanish officers from Spain who replaced the original Spanish conquistadors from Mexico who were more political in alliance-making and who they had treaties of friendship with (due to their common hatred against Muslims, since native Pagan Filipinos fought against the Brunei Sultanate and native Spaniards conquered the Emirate of Granada). The idealistic original pioneers died and were replaced by ignorant royal officers who broke treaties, thus causing the Conspiracy of the Maharlikas among Filipinos, who conspired together with Bruneians and Japanese, yet the failure of the conspiracy caused the royals' exile to the Americas, where they formed communities across the western coasts, chief among which was Guerrero, Mexico, which was later a center of the Mexican war of Independence.
Spanish ocean trade routes and defense
The Spanish crown created a system of convoys of ships (called the flota) to prevent attacks by European privateers. Some isolated attacks on these shipments took place in the Gulf of Mexico and Caribbean Sea by English and Dutch pirates and privateers. One such act of piracy was led by Francis Drake in 1580, and another by Thomas Cavendish in 1587. In one episode, the cities of Huatulco (Oaxaca) and Barra de Navidad in Jalisco Province of México were sacked. However, these maritime routes, both across the Pacific and the Atlantic, were successful in the defensive and logistical role they played in the history of the Spanish Empire. For over three centuries the Spanish Navy escorted the galleon convoys that sailed around the world. Don Lope Díez de Armendáriz, born in Quito, Ecuador, was the first Viceroy of New Spain who was born in the 'New World'. He formed the 'Navy of Barlovento' (Armada de Barlovento), based in Veracruz, to patrol coastal regions and protect the harbors, port towns, and trade ships from pirates and privateers.
Indigenous revolts
After the conquest of central Mexico, there were only two major Indian revolts challenging Spanish rule. In the Mixtón war in 1541, the viceroy Don Antonio de Mendoza led an army against an uprising by Caxcanes. In the 1680 Pueblo revolt, Indians in 24 settlements in New Mexico expelled the Spanish, who left for Texas, an exile lasting a decade. The Chichimeca war lasted over fifty years, 1550–1606, between the Spanish and various indigenous groups of northern New Spain, particularly in silver mining regions and the transportation trunk lines. Non-sedentary or semi-sedentary Northern Indians were difficult to control once they acquired the mobility of the horse. In 1616, the Tepehuan revolted against the Spanish, but it was relatively quickly suppressed. The Tarahumara Indians were in revolt in the mountains of Chihuahua for several years. In 1670 Chichimecas invaded Durango, and the governor, Francisco González, abandoned its defense. The Spanish-Chamorro Wars that began on Guam in 1670 after the Spanish establishment of a physical presence resulted in a series of sieges of the Spanish presidio, the last in 1684.
In the southern area of New Spain, the Tzeltal Maya and other indigenous groups, including the Tzotzil and Chol revolted in 1712. It was a multiethnic revolt sparked by religious issues in several communities. In 1704 viceroy Francisco Fernández de la Cueva suppressed a rebellion of Pima in Nueva Vizcaya.
Bourbon reforms
The Bourbon monarchy embarked upon a far-reaching program to revitalize the economy of its territories, both on the peninsula and its overseas possessions. The crown sought to enhance its control and administrative efficiency, and to decrease the power and privilege of the Roman Catholic Church vis-a-vis the state.
The British capture and occupation of both Manila and Havana in 1762, during the global conflict of the Seven Years' War, meant that the Spanish crown had to rethink its military strategy for defending its possessions. The Spanish crown had engaged with Britain for a number of years in low-intensity warfare, with ports and trade routes harassed by English privateers. The crown strengthened the defenses of Veracruz and San Juan de Ulúa, Jamaica, Cuba, and Florida, but the British sacked ports in the late seventeenth century. Santiago de Cuba (1662), St. Augustine Spanish Florida (1665) and Campeche 1678 and so with the loss of Havana and Manila, Spain realized it needed to take significant steps. The Bourbons created a standing army in New Spain, beginning in 1764, and strengthened defensive infrastructure, such as forts.
The crown sought reliable information about New Spain and dispatched José de Gálvez as Visitador General (inspector general), who observed conditions needing reform, starting in 1765, in order to strengthen crown control over the kingdom.
An important feature of the Bourbon Reforms was that they ended the significant amount of local control that was a characteristic of the bureaucracy under the Habsburgs, especially through the sale of offices. The Bourbons sought a return to the monarchical ideal of having those not directly connected with local elites as administrators, who in theory should be disinterested, staff the higher echelons of regional government. In practice this meant that there was a concerted effort to appoint mostly peninsulares, usually military men with long records of service (as opposed to the Habsburg preference for prelates), who were willing to move around the global empire. The intendancies were one new office that could be staffed with peninsulares, but throughout the 18th century significant gains were made in the numbers of governors-captain generals, audiencia judges and bishops, in addition to other posts, who were Spanish-born.
In 1766, the crown appointed Carlos Francisco de Croix, marqués de Croix as viceroy of New Spain. One of his early tasks was to implement the crown's decision to expel the Jesuits from all its territories, accomplished in 1767. Since the Jesuits had significant power, owning large, well managed haciendas, educating New Spain's elite young men, and as a religious order resistant to crown control, the Jesuits were a major target for the assertion of crown control. Croix closed the religious autos-de-fe of the Holy Office of the Inquisition to public viewing, signaling a shift in the crown's attitude toward religion. Other significant accomplishments under Croix's administration was the founding of the College of Surgery in 1768, part of the crown's push to introduce institutional reforms that regulated professions. The crown was also interested in generating more income for its coffers and Croix instituted the royal lottery in 1769. Croix also initiated improvements in the capital and seat of the viceroyalty, increasing the size of its central park, the Alameda.
Another activist viceroy carrying out reforms was Antonio María de Bucareli y Ursúa, marqués de Valleheroso y conde de Jerena, who served from 1771 to 1779, and died in office. José de Gálvez, now Minister of the Indies following his appointment as Visitor General of New Spain, briefed the newly appointed viceroy about reforms to be implemented. In 1776, a new northern territorial division was established, Commandancy General of the Provincias Internas known as the Provincias Internas (Commandancy General of the Internal Provinces of the North, ). Teodoro de Croix (nephew of the former viceroy) was appointed the first Commander General of the Provincias Internas, independent of the Viceroy of New Spain, to provide better administration for the northern frontier provinces. They included Nueva Vizcaya, Nuevo Santander, Sonora y Sinaloa, Las Californias, Coahuila y Tejas (Coahuila and Texas), and Nuevo México. Bucareli was opposed to Gálvez's plan to implement the new administrative organization of intendancies, which he believed would burden areas with sparse population with excessive costs for the new bureaucracy.
The new Bourbon kings did not split the Viceroyalty of New Spain into smaller administrative units as they did with the Viceroyalty of Peru, carving out the Viceroyalty of Río de la Plata and the Viceroyalty of New Granada, but New Spain was reorganized administratively and elite American-born Spanish men were passed over for high office. The crown also established a standing military, with the aim of defending its overseas territories.
The Spanish Bourbons monarchs' prime innovation introduction of intendancies, an institution emulating that of Bourbon France. They were first introduced on a large scale in New Spain, by the Minister of the Indies José de Gálvez, in the 1770s, who originally envisioned that they would replace the viceregal system (viceroyalty) altogether. With broad powers over tax collection and the public treasury and with a mandate to help foster economic growth over their districts, intendants encroached on the traditional powers of viceroys, governors and local officials, such as the corregidores, which were phased out as intendancies were established. The Crown saw the intendants as a check on these other officers. Over time accommodations were made. For example, after a period of experimentation in which an independent intendant was assigned to Mexico City, the office was thereafter given to the same person who simultaneously held the post of viceroy. Nevertheless, the creation of scores of autonomous intendancies throughout the Viceroyalty, created a great deal of decentralization, and in the Captaincy General of Guatemala, in particular, the intendancy laid the groundwork for the future independent nations of the 19th century. In 1780, Minister of the Indies José de Gálvez sent a royal dispatch to Teodoro de Croix, Commandant General of the Internal Provinces of New Spain (Provincias Internas), asking all subjects to donate money to help the American Revolution. Millions of pesos were given.
The focus on the economy (and the revenues it provided to the royal coffers) was also extended to society at large. Economic associations were promoted, such as the Economic Society of Friends of the Country. Similar "Friends of the Country" economic societies were established throughout the Spanish world, including Cuba and Guatemala.
The crown sent a series of scientific expeditions to its overseas possessions, including the Royal Botanical Expedition to New Spain, led by Martín de Sessé and José Mariano Mociño (1787–1808). Alexander von Humboldt spent a year in New Spain in 1804 on his self-funded scientific expedition to Spanish America. Trained as a mining engineer, Humboldt's observations of silvermining in New Spain were especially important to the crown, which depended on New World silver revenues.
The Bourbon Reforms were not a unified or entirely coherent program, but instead a series of crown initiatives designed to revitalize the economies of its overseas possessions and make administration more efficient and firmly under control of the crown. Record keeping improved and records were more centralized. The bureaucracy was staffed with well-qualified men, most of them peninsular-born Spaniards. The preference for them meant that there was resentment from American-born elite men and their families, who were excluded from holding office. The creation of a military meant that some American Spaniards became officers in local militias, but the ranks were filled with poor, mixed-race men, who resented service and avoided it if possible.
18th-century military conflicts
The first century that saw the Bourbons on the Spanish throne coincided with series of global conflicts that pitted primarily France against Great Britain. Spain, as an ally of Bourbon France, was drawn into these conflicts. In fact part of the motivation for the Bourbon Reforms was the perceived need to prepare the empire administratively, economically, and militarily for what was the next expected war. The Seven Years' War proved to be catalyst for most of the reforms in the overseas possessions, just like the War of the Spanish Succession had been for the reforms on the Peninsula.
In 1720, the Villasur expedition from Santa Fe met and attempted to parley with French- allied Pawnee in what is now Nebraska. Negotiations were unsuccessful, and a battle ensued; the Spanish were badly defeated, with only thirteen managing to return to New Mexico. Although this was a small engagement, it is significant in that it was the deepest penetration of the Spanish into the Great Plains, establishing the limit to Spanish expansion and influence there.
The War of Jenkins' Ear broke out in 1739 between the Spanish and British and was confined to the Caribbean and Georgia. The major action in the War of Jenkins' Ear was a major amphibious attack launched by the British under Admiral Edward Vernon in March 1741 against Cartagena de Indias, one of Spain's major gold-trading ports in the Caribbean (today Colombia). Although this episode is largely forgotten, it ended in a decisive victory for Spain, who managed to prolong its control of the Caribbean and indeed secure the Spanish Main until the 19th century.
Following the French and Indian War/Seven Years' War, the British troops invaded and captured the Spanish cities of Havana in Cuba and Manila in the Philippines in 1762. The Treaty of Paris (1763) gave Spain control over the Louisiana part of New France including New Orleans, creating a Spanish empire that stretched from the Mississippi River to the Pacific Ocean; but Spain also ceded Florida to Great Britain in order to regain Cuba, which the British occupied during the war. Louisiana settlers, hoping to restore the territory to France, in the bloodless Rebellion of 1768 forced the Louisiana Governor Antonio de Ulloa to flee to Spain. The rebellion was crushed in 1769 by the next governor Alejandro O'Reilly, who executed five of the conspirators. The Louisiana territory was to be administered by superiors in Cuba with a governor on site in New Orleans.
The 21 northern missions in present-day California (U.S.) were established along California's El Camino Real from 1769. In an effort to exclude Britain and Russia from the eastern Pacific, King Charles III of Spain sent forth from Mexico a number of expeditions to the Pacific Northwest between 1774 and 1793. Spain's long-held claims and navigation rights were strengthened and a settlement and fort were built in Nootka Sound, Alaska.
Spain entered the American Revolutionary War as an ally of the United States and France in June 1779. From September 1779 to May 1781, Bernardo de Galvez led an army in a campaign along the Gulf Coast against the British. Galvez's army consisted of Spanish regulars from throughout Latin America and a militia which consisted of mostly Acadians along with Creoles, Germans, and Native Americans. Galvez's army engaged and defeated the British in battles fought at Manchac and Baton Rouge, Louisiana, Natchez, Mississippi, Mobile, Alabama, and Pensacola, Florida. The loss of Mobile and Pensacola left the British with no bases along the Gulf Coast. In 1782, forces under Galvez's overall command captured the British naval base at Nassau on New Providence Island in the Bahamas. Galvez was angry that the operation had proceeded against his orders to cancel, and ordered the arrest and imprisonment of Francisco de Miranda, aide-de-camp of Juan Manuel Cajigal, the commander of the expedition. Miranda later ascribed this action on the part of Galvez to jealousy of Cajigal's success.
In the second Treaty of Paris (1783), which ended the American Revolution, Great Britain returned control of Florida back to Spain in exchange for the Bahamas. Spain then had control over the Mississippi River south of 32°30' north latitude, and, in what is known as the Spanish Conspiracy, hoped to gain greater control of Louisiana and all of the west. These hopes ended when Spain was pressured into signing Pinckney's Treaty in 1795. France re-acquired Louisiana from Spain in the secret Treaty of San Ildefonso in 1800. The United States bought the territory from France in the Louisiana Purchase of 1803.
New Spain claimed the entire west coast of North America and therefore considered the Russian fur trading activity in Alaska, which began in the middle to late 18th century, an encroachment and threat. Likewise, the exploration of the northwest coast by Captain James Cook of the British Navy and the subsequent fur trading activities by British ships was considered an encroachment on Spanish territory. To protect and strengthen its claim, New Spain sent a number of expeditions to the Pacific Northwest between 1774 and 1793. In 1789, a naval outpost called Santa Cruz de Nuca (or just Nuca) was established at Friendly Cove in Nootka Sound (now Yuquot), Vancouver Island. It was protected by an artillery land battery called Fort San Miguel. Santa Cruz de Nuca was the northernmost establishment of New Spain. It was the first European colony in what is now the province of British Columbia and the only Spanish settlement in what is now Canada. Santa Cruz de Nuca remained under the control of New Spain until 1795, when it was abandoned under the terms of the third Nootka Convention. Another outpost, intended to replace Santa Cruz de Nuca, was partially built at Neah Bay on the southern side of the Strait of Juan de Fuca in what is now the U.S. state of Washington. Neah Bay was known as Bahía de Núñez Gaona in New Spain, and the outpost there was referred to as "Fuca." It was abandoned, partially finished, in 1792. Its personnel, livestock, cannons, and ammunition were transferred to Nuca.
In 1789, at Santa Cruz de Nuca, a conflict occurred between the Spanish naval officer Esteban José Martínez and the British merchant James Colnett, triggering the Nootka Crisis, which grew into an international incident and the threat of war between Britain and Spain. The first Nootka Convention averted the war but left many specific issues unresolved. Both sides sought to define a northern boundary for New Spain. At Nootka Sound, the diplomatic representative of New Spain, Juan Francisco de la Bodega y Quadra, proposed a boundary at the Strait of Juan de Fuca, but the British representative, George Vancouver refused to accept any boundary north of San Francisco. No agreement could be reached and the northern boundary of New Spain remained unspecified until the Adams–Onís Treaty with the United States (1819). That treaty also ceded Spanish Florida to the United States.
End of the Viceroyalty (1806–1821)
The Third Treaty of San Ildefonso ceded to France the vast territory that Napoleon then sold to the United States in 1803, known as the Louisiana Purchase. The United States obtained Spanish Florida in 1819 in the Adams–Onís Treaty. That treaty also defined a northern border for New Spain, at 42° north latitude (now the northern boundary of the U.S. states of California, Nevada, and Utah).
In the 1821 Declaration of Independence of the Mexican Empire, both Mexico and Central America declared their independence after three centuries of Spanish rule and formed the First Mexican Empire, although Central America quickly rejected the union. After priest Miguel Hidalgo y Costilla's 1810 Grito de Dolores (call for independence), the insurgent army began an eleven-year war. At first, the Criollo class fought against the rebels. But in 1820, a military coup in Spain forced Ferdinand VII to accept the authority of the liberal Spanish Constitution. The specter of liberalism that could undermine the authority and autonomy of the Roman Catholic Church made the Church hierarchy in New Spain view independence in a different light. In an independent nation, the Church anticipated retaining its power. Royalist military officer Agustín de Iturbide proposed uniting with the insurgents with whom he had battled, and gained the alliance of Vicente Guerrero, leader of the insurgents in a region now bearing his name, a region that was populated by immigrants from Africa and the Philippines, crucial among which was the Filipino-Mexican General Isidoro Montes de Oca who impressed Criollo Royalist Itubide into joining forces with Vicente Guerrero by Isidoro Montes De Oca defeating royalist forces three times larger than his, in the name of his leader, Vicente Guerrero. Royal government collapsed in New Spain and the Army of the Three Guarantees marched triumphantly into Mexico City in 1821.
The new Mexican Empire offered the crown to Ferdinand VII or to a member of the Spanish royal family that he would designate. After the refusal of the Spanish monarchy to recognize the independence of Mexico, the (Army of the Three Guarantees), led by Agustín de Iturbide and Vicente Guerrero, cut all political and economic ties with Spain and crowned Iturbide as emperor Agustín of Mexico. Central America was originally planned to be part of the Mexican Empire; but it seceded peacefully in 1823, forming the United Provinces of Central America under the Constitution of 1824.
This left only Cuba and Puerto Rico in the Spanish West Indies, and the Philippines in the Spanish East Indies as part of the Spanish Empire; until their loss to the United States in the Spanish–American War (1898). Before, the Spanish-American War, the Philippines had an almost successful revolt against Spain under the uprising of Andres Novales which were supported by Criollos and Latin Americans who were the Philippines, mainly by the former Latino officers “americanos”, composed mostly of Mexicans with a sprinkling of Creoles and Mestizos from the now independent nations of Colombia, Venezuela, Peru, Chile, Argentina and Costa Rica. went out to start a revolt. In the aftermath, Spain, in order to ensure obedience to the empire, disconnected the Philippines from her Latin-American allies and placed in the Spanish army of the colony, Peninsulars from the mainland which displaced and angered the Latin American and Filipino soldiers who were at the Philippines.
Economy
[[File:8 reales Carolus III 1778 chop.png|thumb|250px|Silver coin minted in New Spain. Silver was its most important export, starting in the 16th century. '8 reales Carlos III - 1778]]
During the era of the conquest, in order to pay off the debts incurred by the conquistadors and their companies, the new Spanish governors awarded their men grants of native tribute and labor, known as encomiendas. In New Spain these grants were modeled after the tribute and corvee labor that the Mexica rulers had demanded from native communities. This system came to signify the oppression and exploitation of natives, although its originators may not have set out with such intent. In short order the upper echelons of patrons and priests in the society lived off the work of the lower classes. Due to some horrifying instances of abuse against the indigenous peoples, Bishop Bartolomé de las Casas suggested bringing black slaves to replace them. Fray Bartolomé later repented when he saw the even worse treatment given to the black slaves.
In colonial Mexico, encomenderos de negros were specialized middlemen during the first half of the seventeenth century. While encomendero (alternatively, encomenderos de indios) generally refers to men granted the labor and tribute of a particular indigenous group in the immediate post-conquest era, encomenderos de negros were Portuguese slave dealers who were permitted to operate in Mexico for the slave trade.
In Peru, the other discovery that perpetuated the system of forced labor, the mit'a, was the enormously rich single silver mine discovered at Potosí, but in New Spain, labor recruitment differed significantly. With the exception of silver mines worked in the Aztec period at Taxco, southwest of Tenochtitlan, the Mexico's mining region was outside the area of dense indigenous settlement. Labor for the mines in the north of Mexico had a workforce of black slave labor and indigenous wage labor, not draft labor. Indigenous who were drawn to the mining areas were from different regions of the center of Mexico, with a few from the north itself. With such diversity they did not have a common ethnic identity or language and rapidly assimilated to Hispanic culture. Although mining was difficult and dangerous, the wages were good, which is what drew the indigenous labor.
The Viceroyalty of New Spain was the principal source of income for Spain in the eighteenth century, with the revival of mining under the Bourbon Reforms. Important mining centers like Zacatecas, Guanajuato, San Luis Potosí and Hidalgo had been established in the sixteenth century and suffered decline for a variety of reasons in the seventeenth century, but silver mining in Mexico out-performed all other Spanish overseas territories in revenues for the royal coffers.
The fast red dye cochineal was an important export in areas such as central Mexico and Oaxaca in terms of revenues to the crown and stimulation of the internal market of New Spain. Cacao and indigo were also important exports for the New Spain, but was used through rather the vice royalties rather than contact with European countries due to piracy, and smuggling. The indigo industry in particular also helped to temporarily unite communities throughout the Kingdom of Guatemala due to the smuggling.
There were two major ports in New Spain, Veracruz the viceroyalty's principal port on the Atlantic, and Acapulco on the Pacific, terminus of the Manila Galleon. In the Philippines Manila near the South China Sea was the main port. The ports were fundamental for overseas trade, stretching a trade route from Asia, through the Manila Galleon to the Spanish mainland.
These were ships that made voyages from the Philippines to Mexico, whose goods were then transported overland from Acapulco to Veracruz and later reshipped from Veracruz to Cádiz in Spain. So then, the ships that set sail from Veracruz were generally loaded with merchandise from the East Indies originating from the commercial centers of the Philippines, plus the precious metals and natural resources of Mexico, Central America, and the Caribbean. During the 16th century, Spain held the equivalent of US$1.5 trillion (1990 terms) in gold and silver received from New Spain.
However, these resources did not translate into development for the Metropolis (mother country) due to Spanish Roman Catholic Monarchy's frequent preoccupation with European wars (enormous amounts of this wealth were spent hiring mercenaries to fight the Protestant Reformation), as well as the incessant decrease in overseas transportation caused by assaults from companies of British buccaneers, Dutch corsairs and pirates of various origin. These companies were initially financed by, at first, by the Amsterdam stock market, the first in history and whose origin is owed precisely to the need for funds to finance pirate expeditions, as later by the London market. The above is what some authors call the "historical process of the transfer of wealth from the south to the north."
Regions of mainland New Spain
In the colonial period, basic patterns of regional development emerged and strengthened. European settlement and institutional life was built in the Mesoamerican heartland of the Aztec Empire in Central Mexico. The South (Oaxaca, Michoacan, Yucatán, and Central America) was a region of dense indigenous settlement of Mesoamerica, but without exploitable resources of interest to Europeans, the area attracted few Europeans, while the indigenous presence remained strong. The North was outside the area of complex indigenous populations, inhabited primarily by nomadic and hostile northern indigenous groups. With the discovery of silver in the north, the Spanish sought to conquer or pacify those peoples in order to exploit the mines and develop enterprises to supply them. Nonetheless, much of northern New Spain had sparse indigenous population and attracted few Europeans. The Spanish crown and later the Republic of Mexico did not effectively exert sovereignty over the region, leaving it vulnerable to the expansionism of the United States in the nineteenth century.
Regional characteristics of colonial Mexico have been the focus of considerable study within the vast scholarship on centers and peripheries. For those based in the vice-regal capital of Mexico City itself, everywhere else were the "provinces." Even in the modern era, "Mexico" for many refers solely to Mexico City, with the pejorative view of anywhere but the capital is a hopeless backwater. "Fuera de México, todo es Cuauhtitlán" ["outside of Mexico City, it's all Podunk"], that is, poor, marginal, and backward, in short, the periphery. The picture is far more complex, however; while the capital is enormously important as the center of power of various kinds (institutional, economic, social), the provinces played a significant role in colonial Mexico. Regions (provinces) developed and thrived to the extent that they were sites of economic production and tied into networks of trade. "Spanish society in the Indies was import-export oriented at the very base and in every aspect," and the development of many regional economies was usually centered on support of that export sector.
Central region
Mexico City, Capital of the Viceroyalty
Mexico City was the center of the Central region, and the hub of New Spain. The development of Mexico City itself is extremely important to the development of New Spain as a whole. It was the seat of the Viceroyalty of New Spain, the Archdiocese of the Catholic Church, the Holy Office of the Inquisition, the merchants' guild (consulado), and home of the most elite families in the Kingdom of New Spain. Mexico City was the single-most populous city, not just in New Spain, but for many years the entire Western Hemisphere, with a high concentration of mixed-race castas.
Veracruz to Mexico City
Significant regional development grew along the main transportation route from the capital east to the port of Veracruz. Alexander von Humboldt called this area "Mesa de Anahuac", which can be defined as the adjacent valleys of Puebla, Mexico, and Toluca, enclosed by high mountains, along with their connections to the Gulf Coast port of Veracruz and the Pacific port of Acapulco, where over half the population of New Spain lived. These valleys were linked trunk lines, or main routes, facilitating the movement of vital goods and people to get to key areas. However, even in this relatively richly endowed region of Mexico, the difficulty of transit of people and goods in the absence of rivers and level terrain remained a major challenge to the economy of New Spain. This challenge persisted during the post-independence years until the late nineteenth-century construction of railroads. In the colonial era and up until the railroads were built in key areas, mule trains were the main mode of transporting goods. Mules were used because unpaved roads and mountainous terrain could not generally accommodate carts.
In the late eighteenth century the crown devoted some resources to the study and remedy the problem of poor roads. The Camino Real (royal road) between the port of Veracruz and the capital had some short sections paved and bridges constructed. The construction was done despite protests from some Indian villages when the infrastructure improvements, which sometimes included rerouting the road through communal lands. The Spanish crown finally decided that road improvement was in the interests of the state for military purposes, as well as for fomenting commerce, agriculture, and industry, but the lack of state involvement in the development of physical infrastructure was to have lasting effects constraining development until the late nineteenth century. Despite some improvements, the roads still made transit difficult, particularly for heavy military equipment.
Although the crown had ambitious plans for both the Toluca and Veracruz portions of the king's highway, actual improvements were limited to a localized network. Even where infrastructure was improved, transit on the Veracruz-Puebla main road had other obstacles, with wolves attacking mule trains, killing animals, and rendering some sacks of foodstuffs unsellable because they were smeared with blood. The north-south Acapulco route remained a mule track through mountainous terrain.
Veracruz, port city and province
Veracruz was the first Spanish settlement founded in what became New Spain, and it endured as the only viable Gulf Coast port, the gateway for Spain to New Spain. The difficult topography around the port affected local development and New Spain as a whole. Going from the port to the central plateau entailed a daunting 2000 meter climb from the narrow tropical coastal plain in just over a hundred kilometers. The narrow, slippery road in the mountain mists was treacherous for mule trains, and in some cases mules were hoisted by ropes. Many tumbled with their cargo to their deaths. Given these transport constraints, only high-value, low-bulk goods continued to be shipped in the transatlantic trade, which stimulated local production of foodstuffs, rough textiles, and other products for a mass market. Although New Spain produced considerable sugar and wheat, these were consumed exclusively in the colony even though there was demand elsewhere. Philadelphia, not New Spain, supplied Cuba with wheat.
The Caribbean port of Veracruz was small, with its hot, pestilential climate not a draw for permanent settlers: its population never topped 10,000. Many Spanish merchants preferred living in the pleasant highland town of Jalapa (1,500 m). For a brief period (1722–76) the town of Jalapa became even more important than Veracruz, after it was granted the right to hold the royal trade fair for New Spain, serving as the entrepot for goods from Asia via Manila Galleon through the port of Acapulco and European goods via the flota (convoy) from the Spanish port of Cádiz. Spaniards also settled in the temperate area of Orizaba, east of the Citlaltepetl volcano. Orizaba varied considerably in elevation from to (the summit of the Citlaltepetl volcano), but "most of the inhabited part is temperate." Some Spaniards lived in semitropical Córdoba, which was founded as a villa in 1618, to serve as a Spanish base against runaway slave (cimarrón) predations on mule trains traveling the route from the port to the capital. Some cimarrón settlements sought autonomy, such as one led by Gaspar Yanga, with whom the crown concluded a treaty leading to the recognition of a largely black town, San Lorenzo de los Negros de Cerralvo, now called the municipality of Yanga.
European diseases immediately affected the multiethnic Indian populations in the Veracruz area and for that reason Spaniards imported black slaves as either an alternative to indigenous labor or its complete replacement in the event of a repetition of the Caribbean die-off. A few Spaniards acquired prime agricultural lands left vacant by the indigenous demographic disaster. Portions of the province could support sugar cultivation and as early as the 1530s sugar production was underway. New Spain's first viceroy, Don Antonio de Mendoza established an hacienda on lands taken from Orizaba.
Indians resisted cultivating sugarcane themselves, preferring to tend their subsistence crops. As in the Caribbean, black slave labor became crucial to the development of sugar estates. During the period 1580–1640 when Spain and Portugal were ruled by the same monarch and Portuguese slave traders had access to Spanish markets, African slaves were imported in large numbers to New Spain and many of them remained in the region of Veracruz. But even when that connection was broken and prices rose, black slaves remained an important component of Córdoba's labor sector even after 1700. Rural estates in Córdoba depended on African slave labor, who were 20% of the population there, a far greater proportion than any other area of New Spain, and greater than even nearby Jalapa.
In 1765 the crown created a monopoly on tobacco, which directly affected agriculture and manufacturing in the Veracruz region. Tobacco was a valuable, high-demand product. Men, women, and even children smoked, something commented on by foreign travelers and depicted in eighteenth-century casta paintings. The crown calculated that tobacco could produce a steady stream of tax revenues by supplying the huge Mexican demand, so the crown limited zones of tobacco cultivation. It also established a small number of manufactories of finished products, and licensed distribution outlets (estanquillos). The crown also set up warehouses to store up to a year's worth of supplies, including paper for cigarettes, for the manufactories. With the establishment of the monopoly, crown revenues increased and there is evidence that despite high prices and expanding rates of poverty, tobacco consumption rose while at the same time, general consumption fell.
In 1787 during the Bourbon Reforms Veracruz became an intendancy, a new administrative unit.
Valley of Puebla
Founded in 1531 as a Spanish settlement, Puebla de los Angeles quickly rose to the status of Mexico's second-most important city. Its location on the main route between the viceregal capital and the port of Veracruz, in a fertile basin with a dense indigenous population, largely not held in encomienda, made Puebla a destination for many later arriving Spaniards. If there had been significant mineral wealth in Puebla, it could have been even more prominent a center for New Spain, but its first century established its importance. In 1786 it became the capital of an intendancy of the same name.
It became the seat of the richest diocese in New Spain in its first century, with the seat of the first diocese, formerly in Tlaxcala, moved there in 1543. Bishop Juan de Palafox asserted the income from the diocese of Puebla as being twice that of the archbishopic of Mexico, due to the tithe income derived from agriculture. In its first hundred years, Puebla was prosperous from wheat farming and other agriculture, as the ample tithe income indicates, plus manufacturing woolen cloth for the domestic market. Merchants, manufacturers, and artisans were important to the city's economic fortunes, but its early prosperity was followed by stagnation and decline in the seventeenth and eighteenth centuries.
The foundation of the town of Puebla was a pragmatic social experiment to settle Spanish immigrants without encomiendas to pursue farming and industry. Puebla was privileged in a number of ways, starting with its status as a Spanish settlement not founded on existing indigenous city-state, but with a significant indigenous population. It was located in a fertile basin on a temperate plateau in the nexus of the key trade triangle of Veracruz–Mexico City–Antequera (Oaxaca). Although there were no encomiendas in Puebla itself, encomenderos with nearby labor grants settled in Puebla. And despite its foundation as a Spanish city, sixteenth-century Puebla had Indians resident in the central core.
Administratively Puebla was far enough away from Mexico City (approximately ) so as not to be under its direct influence. Puebla's Spanish town council (cabildo) had considerable autonomy and was not dominated by encomenderos. The administrative structure of Puebla "may be seen as a subtle expression of royal absolutism, the granting of extensive privileges to a town of commoners, amounting almost to republican self-government, in order to curtail the potential authority of encomenderos and the religious orders, as well as to counterbalance the power of the viceregal capital."
During the "golden century" from its founding in 1531 until the early 1600s, Puebla's agricultural sector flourished, with small-scale Spanish farmers plowing the land for the first time, planting wheat and vaulting Puebla to importance as New Spain's breadbasket, a role assumed by the Bajío (including Querétaro) in the seventeenth century, and Guadalajara in the eighteenth. Puebla's wheat production was the initial element of its prosperity, but it emerged as a manufacturing and commercial center, "serving as the inland port of Mexico's Atlantic trade." Economically, the city received exemptions from the alcabala (sales tax) and almojarifazgo (import/export duties) for its first century (1531–1630), which helped promote commerce.
Puebla built a significant manufacturing sector, mainly in textile production in workshops (obrajes), supplying New Spain and markets as far away as Guatemala and Peru. Transatlantic ties between a particular Spanish town, Brihuega, and Puebla demonstrate the close connection between the two settlements. The take-off for Puebla's manufacturing sector did not simply coincide with immigration from Brihuega but was crucial to "shaping and driving Puebla's economic development, especially in the manufacturing sector." Brihuega immigrants not only came to Mexico with expertise in textile production, but the transplanted briocenses provided capital to create large-scale obrajes. Although obrajes in Brihuega were small-scale enterprises, quite a number of them in Puebla employed up to 100 workers. Supplies of wool, water for fulling mills, and labor (free indigenous, incarcerated Indians, black slaves) were available. Although much of Puebla's textile output was rough cloth, it also produced higher quality dyed cloth with cochineal from Oaxaca and indigo from Guatemala. But by the eighteenth century, Querétaro had displaced Puebla as the mainstay of woolen textile production.
In 1787, Puebla became an intendancy as part of the new administrative structuring of the Bourbon Reforms.
Valley of Mexico
Mexico City dominated the Valley of Mexico, but the valley continued to have dense indigenous populations challenged by growing, increasingly dense Spanish settlement. The Valley of Mexico had many former Indian city-states that became Indian towns in the colonial era. These towns continued to be ruled by indigenous elites under the Spanish crown, with an indigenous governor and a town councils. These Indian towns close to the capital were the most desirable ones for encomenderos to hold and for the friars to evangelize.
The capital was provisioned by the indigenous towns, and its labor was available for enterprises that ultimately created a colonial economy. The gradual drying up of the central lake system created more dry land for farming, but the sixteenth-century population declines allowed Spaniards to expand their acquisition of land. One region that retained strong Indian land holding was the southern fresh water area, with important suppliers of fresh produce to the capital. The area was characterized by intensely cultivated chinampas, man-made extensions of cultivable land into the lake system. These chinampa towns retained a strong indigenous character, and Indians continued to hold the majority of that land, despite its closeness to the Spanish capital. A key example is Xochimilco.
Texcoco in the pre-conquest period was one of the three members of the Aztec Triple Alliance and the cultural center of the empire. It fell on hard times in the colonial period as an economic backwater. Spaniards with any ambition or connections would be lured by the closeness of Mexico City, so that the Spanish presence was minimal and marginal.
Tlaxcala, the major ally of the Spanish against the Aztecs of Tenochtitlan, also became something of a backwater, but like Puebla it did not come under the control of Spanish encomenderos. No elite Spaniards settled there, but like many other Indian towns in the Valley of Mexico, it had an assortment of small-scale merchants, artisans, farmers and ranchers, and textile workshops (obrajes).
North
Since portions of northern New Spain became part of the United States' Southwest region, there has been considerable scholarship on the Spanish borderlands in the north. The motor of the Spanish colonial economy was the extraction of silver. In Bolivia, it was from the single rich mountain of Potosí; but in New Spain, there were two major mining sites, one in Zacatecas, the other in Guanajuato.
The region farther north of the main mining zones attracted few Spanish settlers. Where there were settled indigenous populations, such as in the present-day state of New Mexico and in coastal regions of Baja and Alta California, indigenous culture retained considerable integrity.
Bajío, Mexico's Breadbasket
The Bajío, a rich, fertile lowland just north of central Mexico, was nonetheless a frontier region between the densely populated plateaus and valleys of Mexico's center and south and the harsh northern desert controlled by nomadic Chichimeca. Devoid of settled indigenous populations in the early sixteenth century, the Bajío did not initially attract Spaniards, who were much more interested in exploiting labor and collecting tribute whenever possible. The region did not have indigenous populations that practiced subsistence agriculture. The Bajío developed in the colonial period as a region of commercial agriculture.
The discovery of mining deposits in Zacatecas and Guanajuato in the mid-sixteenth century and later in San Luis Potosí stimulated the Bajío's development to supply the mines with food and livestock. A network of Spanish towns was established in this region of commercial agriculture, with Querétaro also becoming a center of textile production. Although there were no dense indigenous populations or network of settlements, Indians migrated to the Bajío to work as resident employees on the region's haciendas and ranchos or rented land (terrasguerros). From diverse cultural backgrounds and with no sustaining indigenous communities, these indios were quickly hispanized, but largely remained at the bottom of the economic hierarchy. Although Indians migrated willingly to the region, they did so in such small numbers that labor shortages prompted Spanish hacendados to provide incentives to attract workers, especially in the initial boom period of the early seventeenth century. Land owners lent workers money, which could be seen as a perpetual indebtedness, but it can be seen not as coercing Indians to stay but a way estate owners sweetened their terms of employment, beyond their basic wage labor. For example, in 1775 the Spanish administrator of a San Luis Potosí estate "had to scour both Mexico City and the northern towns to find enough blue French linen to satisfy the resident employees." Other types of goods they received on credit were textiles, hats, shoes, candles, meat, beans, and a guaranteed ration of maize. However, where labor was more abundant or market conditions depressed, estate owners paid lower wages. The more sparsely populated northern Bajío tended to pay higher wages than the southern Bajío, which was increasingly integrated in the economy of central Mexico. The credit-based employment system often privileged those holding higher ranked positions on the estate (supervisors, craftsmen, other specialists) who were mostly white, and the estates did not demand repayment.
In the late colonial period, renting complemented estate employment for many non-Indians in more central areas of the Bajío with access to markets. As with hacendados, renters produced for the commercial market. While these Bajío renters could prosper in good times and achieved a level of independence, drought and other disasters made their choice more risky than beneficial.
Many renters retained ties to the estates, diversifying their household's sources of income and level of economic security. In San Luis Potosí, rentals were fewer and estate employment the norm. After a number of years of drought and bad harvests in the first decade of the nineteenth century Hidalgo's 1810 grito appealed more in the Bajío than in San Luis Potosí. In the Bajío estate owners were evicting tenants in favor of renters better able to pay more for land, there was a disruption of previous patterns of mutual benefit between estate owners and renters.
Spanish Borderlands
Areas of northern Mexico were incorporated into the United States in the mid-nineteenth century, following Texas independence and the Mexican–American War (1846–48) and generally known as the "Spanish Borderlands." Scholars in the United States have extensively studied this northern region, which became the states of Texas, New Mexico, Arizona, and California. During the period of Spanish rule, this area was sparsely populated even by indigenous peoples.
The Presidios (forts), pueblos (civilian towns) and the misiones (missions) were the three major agencies employed by the Spanish crown to extend its borders and consolidate its colonial holdings in these territories.
Missions and the Northern Frontier
The town of Albuquerque (present day Albuquerque, New Mexico) was founded in 1706. Other Mexican towns in the region included Paso del Norte (present day Ciudad Juárez), founded in 1667; Santiago de la Monclova in 1689; Panzacola, Tejas in 1681; and San Francisco de Cuéllar (present day city of Chihuahua) in 1709. From 1687, Father Eusebio Francisco Kino, with funding from the Marqués de Villapuente, founded over twenty missions in the Sonoran Desert (in present-day Sonora and Arizona). From 1697, Jesuits established eighteen missions throughout the Baja California Peninsula. Between 1687 and 1700 several missions were founded in Trinidad, but only four survived as Amerindian villages throughout the 18th century. In 1691, explorers and missionaries visited the interior of Texas and came upon a river and Amerindian settlement on 13 June, the feast day of St. Anthony, and named the location and river San Antonio in his honor.
New Mexico
During the term of viceroy Don Luis de Velasco, marqués de Salinas the crown ended the long-running Chichimeca War by making peace with the semi-nomadic Chichimeca indigenous tribes of northern México in 1591. This allowed expansion into the 'Province of New Mexico' or Provincia de Nuevo México. In 1595, Don Juan de Oñate, son of one of the key figures in the silver remining region of Zacatecas, received official permission from the viceroy to explore and conquer New Mexico. As was the pattern of such expeditions, the leader assumed the greatest risk but would reap the largest rewards, so that Oñate would become capitán general of New Mexico and had the authority to distribute rewards to those in the expedition. Oñate pioneered 'The Royal Road of the Interior Land' or El Camino Real de Tierra Adentro between Mexico City and the Tewa village of Ohkay Owingeh, or San Juan Pueblo. He also founded the Spanish settlement of San Gabriel de Yungue-Ouinge on the Rio Grande near the Native American Pueblo, located just north of the present day city of Española, New Mexico. Oñate eventually learned that New Mexico, while it had a settled indigenous population, had little arable land, had no silver mines, and possessed few other resources to exploit that would merit large scale colonization. He resigned as governor in 1607 and left New Mexico, having lost much of his personal wealth on the enterprise.
In 1610, Pedro de Peralta, a later governor of the Province of New Mexico, established the settlement of Santa Fe near the southern end of the Sangre de Cristo mountain range. Missions were established to convert the indigenous peoples and manage the agricultural industry. The territory's indigenous population resented the Spanish forced conversion to Christianity and suppression of their religion, and the imposition of encomienda system of forced labor. The unrest led to the Pueblo Revolt in 1680, expelling the Spanish, who retreated to Paso del Norte (modern-day Ciudad Juárez.) After the return of the Spanish in 1692, the final resolution included a marked reduction of Spanish efforts to eradicate native culture and religion, the issuing of substantial communal land grants to each Pueblo, and a public defender of their rights and for their legal cases in Spanish courts. In 1776 the New Mexico came under the new Provincias Internas jurisdiction. In the late 18th century the Spanish land grant encouraged the settlement by individuals of large land parcels outside Mission and Pueblo boundaries, many of which became ranchos.
California
In 1602, Sebastián Vizcaíno, the first Spanish presence in the 'New California' (Nueva California) region of the frontier Las Californias province since Cabrillo in 1542, sailed as far north up the Pacific Coast as present-day Oregon, and named California coastal features from San Diego to as far north as the Bay of Monterrey.
Not until the eighteenth century was California of much interest to the Spanish crown, since it had no known rich mineral deposits or indigenous populations sufficiently organized to render tribute and do labor for Spaniards. The discovery of huge deposits of gold in the Sierra Nevada foothills did not come until after the U.S. had incorporated California following the Mexican–American War (1846–48).
By the middle of the 1700s, the Catholic order of Jesuits had established a number of missions on the Baja (lower) California peninsula. Then, in 1767, King Charles III ordered all Jesuits expelled from all Spanish possessions, including New Spain. New Spain's Visitador General José de Gálvez replaced them with the Dominican Order in Baja California, and the Franciscans were chosen to establish new northern missions in Alta (upper) California.
In 1768, Gálvez received the following orders: "Occupy and fortify San Diego and Monterey for God and the King of Spain." The Spanish colonization there, with far fewer known natural resources and less cultural development than Mexico or Peru, was to combine establishing a presence for defense of the territory with a perceived responsibility to convert the indigenous people to Christianity.
The method used to "occupy and fortify" was the established Spanish colonial system: missions (misiones, between 1769 and 1833 twenty-one missions were established) aimed at converting the Native Californians to Christianity, forts (presidios, four total) to protect the missionaries, and secular municipalities (pueblos, three total). Due to the region's great distance from supplies and support in México, the system had to be largely self-sufficient. As a result, the colonial population of California remained small, widely scattered and near the coast.
In 1776, the north-western frontier areas came under the administration of the new 'Commandancy General of the Internal Provinces of the North' (Provincias Internas), designed to streamline administration and invigorate growth. The crown created two new provincial governments from the former Las Californias in 1804; the southern peninsula became Baja California, and the ill-defined northern mainland frontier area became Alta California.
Once missions and protective presidios were established in an area, large land grants encouraged settlement and establishment of California ranchos. The Spanish system of land grants was not very successful, however, because the grants were merely royal concessions—not actual land ownership. Under later Mexican rule, land grants conveyed ownership, and were more successful at promoting settlement.
Rancho activities centered on cattle-raising; many grantees emulated the Dons of Spain, with cattle, horses and sheep the source of wealth. The work was usually done by Native Americans, sometimes displaced and/or relocated from their villages. Native-born descendants of the resident Spanish-heritage rancho grantees, soldiers, servants, merchants, craftsmen and others became the Californios. Many of the less-affluent men took native wives, and many daughters married later English, French and American settlers.
After the Mexican War of Independence (1821) and subsequent secularization ("disestablishment") of the missions (1834), Mexican land grant transactions increased the spread of the rancho system. The land grants and ranchos established mapping and land-ownership patterns that are still recognizable in present-day California and New Mexico.
South
Yucatán
The Yucatán peninsula can be seen as a cul-de-sac, and it does indeed have unique features, but it also has strong similarities to other areas in the South. The Yucatán peninsula extends into the Gulf of Mexico and was connected to Caribbean trade routes and Mexico City, far more than some other southern regions, such as Oaxaca. There was three main Spanish settlements, the inland city of Mérida, where Spanish civil and religious officials had their headquarters and where the many Spaniards in the province lived. The villa of Campeche was the peninsula's port, the key gateway for the whole region. A merchant group developed and expanded dramatically as trade flourished during the seventeenth century. Although that period was once characterized as New Spain's "century of depression," for Yucatán this was certainly not the case, with sustained growth from the early seventeenth century to the end of the colonial period.
With dense indigenous Maya populations, Yucatán's encomienda system was established early and persisted far longer than in central Mexico, since fewer Spaniards migrated to the region than in the center. Although Yucatán was a more peripheral area to the colony, since it lacked rich mining areas and no agricultural or other export product, it did have a complex of Spanish settlement, with a whole range of social types in the main settlements of Mérida and the villas of Campeche and Valladolid. There was an important sector of mixed-race "castas", some of whom were fully at home in both the indigenous and Hispanic worlds. Blacks were an important component of Yucatecan society. The largest population in the province was indigenous Maya, who lived in their communities, but which were in contact with the Hispanic sphere via labor demands and commerce.
In Yucatán, Spanish rule was largely indirect, allowing these communities considerable political and cultural autonomy. The Maya community, the cah, was the means by which indigenous cultural integrity was maintained. In the economic sphere, unlike many other regions and ethnic groups in Mesoamerica, the Yucatec Maya did not have a pre-conquest network of regular markets to exchange different types of food and craft goods. Perhaps because the peninsula was uniform in its ecosystem local niche production did not develop. Production of cotton textiles, largely by Maya women, helped pay households' tribute obligations, but basic crops were the basis of the economy. The cah retained considerable land under the control of religious brotherhoods or confraternities (cofradías), the device by which Maya communities avoided colonial officials, the clergy, or even indigenous rulers (gobernadores) from diverting of community revenues in their cajas de comunidad (literally community-owned chests that had locks and keys). Cofradías were traditionally lay pious organizations and burial societies, but in Yucatán they became significant holders of land, a source of revenue for pious purposes kept under cah control. "[I]n Yucatán the cofradía in its modified form was the community." Local Spanish clergy had no reason to object to the arrangement since much of the revenue went for payment for masses or other spiritual matters controlled by the priest.
A limiting factor in Yucatán's economy was the poorness of the limestone soil, which could only support crops for two to three years with land cleared through slash and burn agriculture. Access to water was a limiting factor on agriculture, with the limestone escarpment giving way in water filled sinkholes (locally called cenotes), but rivers and streams were generally absent on the peninsula. Individuals had rights to land so long as they cleared and tilled them and when the soil was exhausted, they repeated the process. In general, the Indians lived in a dispersed pattern, which Spanish congregación or forced resettlement attempted to alter. Collective labor cultivated the confraternities' lands, which included raising the traditional maize, beans, and cotton. But confraternities also later pursued cattle ranching, as well as mule and horse breeding, depending on the local situation. There is evidence that cofradías in southern Campeche were involved in inter-regional trade in cacao as well as cattle ranching. Although generally the revenues from crops and animals were devoted to expenses in the spiritual sphere, cofradías' cattle were used for direct aid to community members during droughts, stabilizing the community's food supply.
In the seventeenth century, patterns shifted in Yucatán and Tabasco, as the English took territory the Spanish claimed but did not control, especially what became British Honduras (now Belize) and in Laguna de Términos (Isla del Carmen) where they cut logwood. In 1716–17 viceroy of New Spain organized a sufficient ships to expel the foreigners, where the crown subsequently built a fortress at Isla del Carmen. But the British held onto their territory in the eastern portion of the peninsula into the twentieth century. In the nineteenth century, the enclave supplied guns to the rebellious Maya in the Caste War of Yucatan.
Valley of Oaxaca
Since Oaxaca was lacking in mineral deposits and it had an abundant sedentary indigenous population, its development was notable for the lack of European or mixed-race population, lack of large-scale Spanish haciendas, and the survival of indigenous communities. These communities retained their land, indigenous languages, and distinct ethnic identities. Antequera (now Oaxaca City) was a Spanish settlement founded in 1529, but the rest of Oaxaca consisted of indigenous towns. Despite its remoteness from Mexico City, "throughout the colonial era, Oaxaca was one of Mexico's most prosperous provinces." In the eighteenth century, the value of crown offices (alcalde mayor or corregidor) were the highest for two Oaxaca jurisdictions, with Jicayan and Villa Alta each worth 7,500 pesos, Cuicatlan-Papalotipac, 4,500; Teposcolula and Chichicapa, each 4,200 pesos.
The most important commodity for Oaxaca was cochineal red dye. Cochineal's commodity chain is an interesting one, with indigenous peasants in the remote areas of Oaxaca ultimately linked to Amsterdam and London commodity exchanges and the European production of luxury cloth. The most extensive scholarly work on Oaxaca's eighteenth-century economy deals with the nexus between the local crown officials (alcaldes mayores), merchant investors (aviadores), the repartimiento (forced labor), and indigenous products, particularly cochineal. The rich, color-fast red dye produced from insects, was harvested from nopal cacti. Cochineal was a high-value, low-volume product that became the second-most valuable Mexican export after silver. Although it could be produced elsewhere in central and southern Mexico, its main region of production was Oaxaca. For the indigenous in Oaxaca, cochineal was the only one "with which the [tributaries] maintain themselves and pay their debts" but it also had other advantages for them. Producing cochineal was time-consuming labor, but it was not particularly difficult and could be done by the elderly, women, and children. It was also important to households and communities because it initially did not require the indigenous to displace their existing crops or migrate elsewhere.
Although the repartimiento has historically been seen as an imposition on the indigenous, forcing them into economic relations they would rather have avoided and maintained by force, recent work on eighteenth-century Oaxaca analyzes the nexus of crown officials (the alcaldes mayores) and Spanish merchants, and indigenous via the repartimiento. cash loaned by local crown officials (the alcalde mayor and his teniente), usually to individual Indians but sometimes to communities, in exchange for a fixed amount of a good (cochineal or cotton mantles) at a later date. Indigenous elites were an integral part of the repartimiento, often being recipients of large extensions of credit. As authority figures in their community, they were in a good position to collect on the debt, the most risky part of the business from the Spanish point of view.
Tehuantepec
The Isthmus of Tehuantepec region of Oaxaca was strategically important for its short transit between the Gulf Coast and the Pacific, facilitating both overland and sea trade. The province of Tehuantepec was the Pacific side of the isthmus and the headwaters of the Coatzacoalcos River. Hernán Cortés acquired holdings for his entailed estate including Huatulco, once the main Pacific Coast port before Acapulco replaced it in 1563.
Gold mining was an early draw for Spaniards, who directed indigenous labor to its extraction, but did not continue beyond the mid-sixteenth century. Over the long run, ranching and commerce were the most important economic activities, with the settlement of Tehuantepec becoming the hub. The region's history can be divided into three distinct periods, an initial period of engagement with Spanish colonial rule to 1563, during which there was a working relationship with the Zapotec ruling line and the establishment of Cortés's economic enterprises. This early period came to a close with the death of the last native king in 1562 and the escheatment of Cortés's Tehuantepec encomiendas to the crown in 1563. The second period of approximately a century (1563–1660) saw the decline of the indigenous entailed estate (cacicazgo) and indigenous political power and development of the colonial economy and imposition of Spanish political and religious structures. The final period is the maturation of these structures (1660–1750). The 1660 rebellion can be a dividing line between the two later periods.
The Villa of Tehuantepec, the largest settlement on the isthmus, was an important prehispanic Zapotec trade and religious center, which was not under the jurisdiction of the Aztecs. The early colonial history of Tehuantepec and the larger province was dominated by Cortés and the Marquesado, but the crown realized the importance of the area and concluded an agreement in 1563 with the second Marqués by which the crown took control of the Tehuantepec encomienda. The Marquesado continued to have major private holdings in the province. The Villa of Tehuantepec became a center of Spanish and mixed-race settlement, crown administration, and trade.
The Cortés haciendas in Tehuantepec were key components of the province's economy, and they were directly linked to other Marquesado enterprises in greater Mexico in an integrated fashion. The Dominicans also had significant holdings in Tehuantepec, but there has been little research on these. However important the Marquesado and the Dominican enterprises were, there were also other economic players in the region, including individual Spaniards as well as existing indigenous communities. Ranching emerged as the dominant rural enterprise in most of Tehuantepec with a ranching boom in the period 1580–1640. Since Tehuantepec experienced significant indigenous population loss in the sixteenth century conforming to the general pattern, ranching made possible for Spaniards to thrive in Tehuantepec because ranching was not dependent on significant amounts of indigenous labor.
The most detailed economic records for the region are of the Marquesado's ranching haciendas, which produced draft animals (horses, mules, burros, and oxen) and sheep and goats, for meat and wool. Cattle ranching for meat, tallow, and leather were also important. Tallow for candles used in churches and residences and leather used in a variety of ways (saddles, other tack, boots, furniture, machinery) were significant items in the larger colonial economy, finding markets well beyond Tehuantepec. Since the Marquesado operated as an integrated enterprise, draft animals were used in other holdings for transport, agriculture, and mining in Oaxaca, Morelos, Toluca, and Mexico City as well as sold. Raised in Tehuantepec, the animals were driven to other Marquesado holdings for use and distribution.
Although colonial population decline affected the indigenous in Tehuantepec, their communities remained important in the colonial era and remain distinctly Indian to the current era. There were differences in the three distinct linguistic and ethnic groups in colonial Tehuantepec, the Zapotec, the Zoque, and the Huave. The Zapotecs concluded an alliance with the Spaniards at contact, and they had already expanded their territory into Zoque and Huave regions.
Under Spanish rule, the Zapotecs not only survived, but flourished, unlike the other two. They continued to pursue agriculture, some of it irrigated, which was not disrupted by the growing ranching economy. Generally Zapotec elites protected their communities from Spanish incursions and community cohesion remained strong as shown in members' performance of regular community service for social ends. Zapotec elites engaged in the market economy early on, which undermined to an extent the bonds between commoners and elites who colluded with the Spanish. In contrast to the Zapotecs, the Zoque generally declined as a group during the ranching boom, with interloping animals eating their maize crops. Zoque response was to take up being vaqueros themselves. They had access to the trade to Guatemala. Of the three indigenous groups, the Huave were the most isolated from the Spanish ranching economy and labor demands. With little arable or grazing land, they exploited the lagoons of the Pacific coast, using shore and beach resources. They traded dried shrimp and fish, as well as purple dye from shells to Oaxaca, likely acquiring foodstuffs that they were unable to cultivate themselves.
Not well documented is the number of African slaves and their descendants, who were artisans in urban areas and did hard manual labor in rural areas. In a pattern recognizable elsewhere, coastal populations were mainly African, including an unknown number of cimarrón (runaway slave) settlements, while inland the indigenous communities were more prominent. On the Cortés haciendas, blacks and mulattoes were essential to the profitability of the enterprises.
In general, Tehuantepec was not a site of major historical events, but in 1660–61, there was a significant rebellion stemming from increased repartimiento Spanish demands.
Central America
With the growth of a sufficient Spanish population and the crown's desire to better govern the area, it established the Captaincy General of Guatemala, which had primary jurisdiction over what are now Guatemala, El Salvador, Honduras, Nicaragua, and Costa Rica. The region was diverse, and outlying provinces were resentful for elites in capital of Antigua Guatemala, destroyed by an earthquake in 1773. There was a high court Audiencia in the Kingdom of Guatemala. Given the region's distance from major centers of power in New Spain and Spain itself, local strongmen in the early were only nominally subject to royal authority. The indigenous population was very large in comparison to the Spanish, and there were relatively few Africans. Spaniards continued to employ forced labor in the region starting with the conquest era and exact tribute from the indigenous. Compared to the mining areas of New Spain's North, this region was generally poor in mineral resources, although Honduras had a brief boom in gold mining, and in the colonial period had little potential to develop an export product, except for cacao and the blue dye indigo.
Cacao had been cultivated in the prehispanic period. Orchards of cacao trees, which took a number of years to come to maturity and produce fruit. Cacao boomed in the late sixteenth century, and then was displaced by indigo as the most important export product. Indigo, like cacao, was native to the region, and the indigenous peoples gathered wild indigo, used for dying cloth and as a trade good. After the arrival of the Spanish, they domesticated indigo and created plantations for its cultivation in Yucatan, El Salvador, and Guatemala. The indigo industry thrived, since there was high demand in Europe for a high quality, color-fast blue dye. In the region, cultivation and processing was done by indigenous workers, but the owners of plantations, añileros, were Spanish. It was a dangerous work environment, with toxins present in the indigo plants that sickened and sometimes killed workers. It was profitable, especially following the Bourbon Reforms, which allowed trade within the Spanish empire. In the late eighteenth century, indigo growers organized in a trade organization, the Consulado de Comercio. There were regions that were not subjugated to Spanish rule, such as the Petén and the Mosquito Coast, and the English took advantage of weak Spanish control to establish a commercial presence on the Gulf Coast, later seizing Belize. An American-born Spanish elite (criollos) accumulated land and built fortunes on wheat, sugar, and cattle, all of which were consumed within the region.
Demographics
The role of epidemics
Spanish settlers brought to the American continent smallpox, measles, typhoid fever, and other infectious diseases. Most of the Spanish settlers had developed an immunity to these diseases from childhood, but the indigenous peoples lacked the needed antibodies since these diseases were totally alien to the native population at the time. There were at least three separate, major epidemics that decimated the population: smallpox (1520 to 1521), measles (1545 to 1548) and typhus (1576 to 1581). In the course of the 16th century, the native population in Mexico went from an estimated pre-Columbian population of 8 to 20 million to less than two million. Therefore, at the start of the 17th century, continental New Spain was a depopulated country with abandoned cities and maize fields. These diseases would not affect the Philippines in the same way because the diseases were already present in the country; Pre-Hispanic Filipinos had contact with other foreign nationalities before the arrival of the Spaniards.
Population in early 1800s
While different intendencies would perform censuses to get a detailed insight in regards to its inhabitants (namely occupation, number of persons per household, ethnicity etc.), it was not until 1793 that the results of the first ever national census would be published. The census is also known as the "Revillagigedo census" because its creation was ordered by the Count of the same name. Most of the census' original datasets have reportedly been lost; thus most of what is known about it nowadays comes from essays and field investigations made by academics who had access to the census data and used it as reference for their works, such as Prussian geographer Alexander von Humboldt. Each author gives different estimations for the total population, ranging from 3,799,561 to 6,122,354 (more recent data suggests that the actual population of New Spain in 1810 was closer to 5 or 5.5 million individuals) as well as the ethnic composition in the country although there is not much variation, with Europeans ranging from 18% to 22% of New Spain's population, Mestizos ranging from 21% to 25%, Indians ranging from 51% to 61% and Africans being between 6,000 and 10,000. It is concluded then, that across nearly three centuries of colonization, the population growth trends of whites and Mestizos were even, while the total percentage of the indigenous population decreased at a rate of 13%–17% per century. The authors assert that rather than whites and Mestizos having higher birthrates, the reason for the indigenous population's numbers decreasing lies on them suffering of higher mortality rates, due to living in remote locations rather than in cities and towns founded by the Spanish colonists or being at war with them. It is also for these reasons that the number of Indigenous Mexicans presents the greater variation range between publications, as in cases their numbers in a given location were estimated rather than counted, leading to possible overestimations in some provinces and possible underestimations in others.
~Europeans are included within the Mestizo category.
Regardless of the possible imprecision related to the counting of Indigenous peoples living outside of the colonized areas, the effort that New Spain's authorities put on considering them as subjects is worth mentioning, as censuses made by other colonial or post-colonial countries did not consider American Indians to be citizens/subjects, as example the censuses made by the Viceroyalty of the Río de la Plata would only count the inhabitants of the colonized settlements. Other example would be the censuses made by the United States, that did not include Indigenous peoples living among the general population until 1860, and indigenous peoples as a whole until 1900.
Once New Spain achieved its independence, the legal basis of the Colonial caste system was abolished and mentions of a person's caste in official documents were also abandoned, which led to the exclusion of racial classification in the censuses to come and difficulted to keep track of the demographic development of each ethnicity that lived in the country. More than a century would pass for Mexico to conduct a new census on which a person's race was taken into account, in 1921, but even then, due to it showing huge inconsistencies regarding other official registers as well as its historic context, modern investigators have deemed it inaccurate. Almost a century after the aforementioned census was made, Mexico's government has begun to conduct ethno-racial surveys again, with its results suggesting that the population growth trends for each major ethnic group haven't changed significantly since the 1793 census was taken.
Culture, art, and architecture
The capital of Viceroyalty of New Spain, Mexico City, was one of the principal centers of European cultural expansion in the Americas. Some of the most important early buildings in New Spain were churches and other religious architecture. Civil architecture included the viceregal palace, now the National Palace, and the Mexico City town council (cabildo), both located on the main square in the capital.
The first printing press in the New World was brought to Mexico in 1539, by printer Juan Pablos (Giovanni Paoli). The first book printed in Mexico was entitled "La escala espiritual de San Juan Clímaco". In 1568, Bernal Díaz del Castillo finished La Historia Verdadera de la Conquista de la Nueva España. Figures such as Sor Juana Inés de la Cruz, Juan Ruiz de Alarcón, and don Carlos de Sigüenza y Góngora, stand out as some of the viceroyalty's most notable contributors to Spanish Literature. In 1693, Sigüenza y Góngora published El Mercurio Volante, the first newspaper in New Spain.
Architects Pedro Martínez Vázquez and Lorenzo Rodriguez produced some visually frenetic architecture known as Mexican Churrigueresque in the capital, Ocotlan, Puebla and some remote silver-mining towns. Composers including Manuel de Zumaya, Juan Gutiérrez de Padilla, and Antonio de Salazar were active from the early 1500s through the Baroque period of music.
See also
Criollo people
Economic history of Mexico
Filipino immigration to Mexico
Governor-General of the Philippines
Historiography of Colonial Spanish America
History of democracy in Mexico
History of Honduras
Index of Mexico-related articles
List of governors in the Viceroyalty of New Spain
List of viceroys of New Spain
Mexican settlement in the Philippines
Spanish American Enlightenment
Notes
References
Bibliography
General histories
|
Meyer, Michael C., William L. Sherman, and Susan M. Deeds. (2014) The Course of Mexican History Tenth edition, esp. chapters 6-15. New York: Oxford University Press.
More specialized works
Historiography
Hanke, Lewis. Do the Americas Have a Common History? A Critique of the Bolton Theory (1964)
Hurtado, Albert L. "Bolton and Turner: The Borderlands and American Exceptionalism." Western Historical Quarterly 44#1 (2013): 4–20. online
Hurtado, Albert L. Herbert Eugene Bolton: Historian of the American Borderlands (University of California Press; 2012)
Van Young, Eric. "Two Decades of Anglophone Historical Writing on Colonial Mexico: Continuity and Change since 1980". Mexican Studies/Estudios Mexicanos. (2004) vol. 20, No. 2 (summer), 275-326.
Reference worksEncyclopedia of Mexico. 2 vols. (1997) Chicago.Encyclopedia of Latin American History and Culture. 5 vols. (1996) New York.
Further reading
Bakewell, P.J. A History of Latin America (Oxford U.P., 1997)
Bethell, Leslie, ed. The Cambridge History of Latin America (Vols. 1–2. Cambridge UP, 1984)
Cañeque, Alejandro. "The Political and Institutional History of Colonial Spanish America" History Compass (April 2013) 114 pp 280–291,
Collier, Simon. From Cortes to Castro: An Introduction to the History of Latin America, 1492–1973 (1974)
Gibson, Charles. The Aztecs Under Spanish Rule: A History of the Indians of the Valley of Mexico, 1519–1810. (Stanford University Press 1964).
Lockhart, James. The Nahuas After the Conquest (Stanford University Press)
Muldoon, James. The Americas in The Spanish World Order (1994)
Parry, J.H. The Spanish Seaborne Empire (1974)
Parry, J.H. The Spanish Theory of Empire in the Sixteenth Century (1974)
Stein, Barbara H., and Stanley J. Stein. Crisis in an Atlantic Empire: Spain and New Spain, 1808–1810 (Johns Hopkins University Press; 2014) 808 pages.
Leibsohn, Dana, and Barbara E. Mundy, Vistas: Visual Culture in Spanish America, 1520–1820. http://www.fordham.edu/vistas, 2015.
External links
MEXICO'S COLONIAL ERA—PART I: The Settlement of New Spain at mexconnect.com
Index to the DeWitt Colony Region under New Spain at Texas A&M University
1492 – Middle America at ibiblio.org the public's library and digital archive''
Encyclopædia Britannica: Hispanic Heritage in The Americas
Map of the Border of the King's Dominion in the Northern America is a map by José de Urrútia and Nicolas de la Fora.
Former colonies in North America
Former countries of the United States
Former Spanish colonies
Viceroyalties of the Spanish Empire
.
.
Spain
History of North America
01
Latin American history
Pre-statehood history of Arizona
Pre-statehood history of New Mexico
.
.
.
.
States and territories established in 1519
States and territories established in 1535
1519 in North America
1510s establishments in Mexico
1520s establishments in Mexico
States and territories disestablished in 1821
1821 in North America
1820s disestablishments in Mexico
Titles of nobility in the Americas
Spanish-speaking countries and territories
Spanish exploration in the Age of Discovery
Former regions and territories of the United States |
null | null | Bombing of Dresden in World War II | eng_Latn | The bombing of Dresden was a British-American aerial bombing attack on the city of Dresden, the capital of the German state of Saxony, during World War II. In four raids between 13 and 15 February 1945, 772 heavy bombers of the British Royal Air Force (RAF) and 527 of the United States Army Air Forces (USAAF) dropped more than 3,900 tons of high-explosive bombs and incendiary devices on the city. The bombing and the resulting firestorm destroyed more than of the city centre. An estimated 22,700 to 25,000 people were killed. Three more USAAF air raids followed, two occurring on 2 March aimed at the city's railway marshalling yard and one smaller raid on 17 April aimed at industrial areas.
Immediate German propaganda claims following the attacks and postwar discussions of whether the attacks were justified have led to the bombing becoming one of the moral causes célèbres of the war. A 1953 United States Air Force report defended the operation as the justified bombing of a strategic target, which they noted was a major rail transport and communication centre, housing 110 factories and 50,000 workers in support of the German war effort.
Several researchers claim that not all of the communications infrastructure, such as the bridges, were targeted, nor were the extensive industrial areas which were located outside the city centre. Critics of the bombing have asserted that Dresden was a cultural landmark while downplaying its strategic significance, and claim that the attacks were indiscriminate area bombing and not proportionate to the military gains. Some have claimed that the raid constituted a war crime. Some, mostly in the German far-right, refer to the bombing as a mass murder, calling it "Dresden's Holocaust of bombs".
In the decades since the war, large variations in the claimed death toll have fuelled the controversy, though the numbers themselves are no longer a major point of contention among historians. In March 1945, the German government ordered its press to publish a falsified casualty figure of 200,000 for the Dresden raids, and death tolls as high as 500,000 have been claimed. The city authorities at the time estimated up to 25,000 victims, a figure that subsequent investigations supported, including a 2010 study commissioned by the city council. One of the main authors responsible for inflated figures being disseminated in the West was Holocaust denier David Irving, who subsequently announced that he had discovered that the documentation he had worked from had been forged, and the real figures supported the 25,000 number.
Background
Early in 1945, the German offensive known as the Battle of the Bulge had been exhausted, as was the Luftwaffe's disastrous New Year's Day attack involving elements of 11 combat wings of its day fighter force. The Red Army had launched its Silesian Offensives into pre-war German territory. The German army was retreating on all fronts, but still resisting strongly. On 8 February 1945, the Red Army crossed the Oder River, with positions just from Berlin. A special British Joint Intelligence Subcommittee report, German Strategy and Capacity to Resist, prepared for Winston Churchill's eyes only, predicted that Germany might collapse as early as mid-April if the Soviets overran its eastern defences. Alternatively, the report warned that the Germans might hold out until November if they could prevent the Soviets from taking Silesia. Hence any assistance to the Soviets on the Eastern Front could shorten the war.
A large scale aerial attack on Berlin and other eastern cities was examined under the code name Operation Thunderclap in mid-1944, but was shelved on 16 August. This was later reexamined, and the decision made for a more limited operation. The Soviet Army continued its push towards the Reich despite severe losses, which they sought to minimize in the final phase of the war. On 5 January 1945, two North American B-25 Mitchell bombers dropped 300,000 leaflets over Dresden with the "Appeal of 50 German generals to the German army and people".
On 22 January 1945, the RAF director of bomber operations, Air Commodore Sydney Bufton, sent Deputy Chief of the Air Staff Air Marshal Sir Norman Bottomley a minute suggesting that if Thunderclap was timed so that it appeared to be a coordinated air attack to aid the current Soviet offensive then the effect of the bombing on German morale would be increased. On 25 January, the Joint Intelligence Committee supported the idea, as it tied in with the Ultra-based intelligence that dozens of German divisions deployed in the west were moving to reinforce the Eastern Front, and that interdiction of these troop movements should be a "high priority." Air Chief Marshal Sir Arthur Harris, AOC-in-C Bomber Command, nicknamed "Bomber Harris", and known as an ardent supporter of area bombing, was asked for his view, and proposed a simultaneous attack on Chemnitz, Leipzig and Dresden. That evening Churchill asked the Secretary of State for Air, Sir Archibald Sinclair, what plans had been drawn up to carry out these proposals. He passed on the request to Marshal of the Royal Air Force Sir Charles Portal, the Chief of the Air Staff, who answered, "We should use available effort in one big attack on Berlin and attacks on Dresden, Leipzig, and Chemnitz, or any other cities where a severe blitz will not only cause confusion in the evacuation from the East, but will also hamper the movement of troops from the West." He mentioned that aircraft diverted to such raids should not be taken away from the current primary tasks of destroying oil production facilities, jet aircraft factories, and submarine yards.
Churchill was not satisfied with this answer and on 26 January pressed Sinclair for a plan of operations: "I asked [last night] whether Berlin, and no doubt other large cities in east Germany, should not now be considered especially attractive targets ... Pray report to me tomorrow what is going to be done".
In response to Churchill's inquiry, Sinclair approached Bottomley, who asked Harris to undertake attacks on Berlin, Dresden, Leipzig, and Chemnitz as soon as moonlight and weather permitted, "with the particular object of exploiting the confused conditions which are likely to exist in the above mentioned cities during the successful Russian advance". This allowed Sinclair to inform Churchill on 27 January of the Air Staff's agreement that, "subject to the overriding claims" on other targets under the Pointblank Directive, strikes against communications in these cities to disrupt civilian evacuation from the east and troop movement from the west would be made.
On 31 January, Bottomley sent Portal a message saying a heavy attack on Dresden and other cities "will cause great confusion in civilian evacuation from the east and hamper movement of reinforcements from other fronts". British historian Frederick Taylor mentions a further memo sent to the Chiefs of Staff Committee by Air Marshal Sir Douglas Evill on 1 February, in which Evill states interfering with mass civilian movements was a major, even key, factor in the decision to bomb the city centre. Attacks there, where main railway junctions, telephone systems, city administration and utilities were, would result in "chaos". Ostensibly, Britain had learned this after the Coventry Blitz, when loss of this crucial infrastructure had supposedly longer-lasting effects than attacks on war plants.
During the Yalta Conference on 4 February, the Deputy Chief of the Soviet General Staff, General Aleksei Antonov, raised the issue of hampering the reinforcement of German troops from the western front by paralyzing the junctions of Berlin and Leipzig with aerial bombardment. In response, Portal, who was in Yalta, asked Bottomley to send him a list of objectives to discuss with the Soviets. Bottomley's list included oil plants, tank and aircraft factories and the cities of Berlin and Dresden. However according to Richard Overy, the discussion with the Soviet Chief of Staff, Aleksei Antonov, recorded in the minutes, only mentions the bombing of Berlin and Leipzig. The bombing of Dresden was a Western plan, but the Soviets were told in advance about the operation.
Military and industrial profile
According to the RAF at the time, Dresden was Germany's seventh-largest city and the largest remaining unbombed built-up area. Taylor writes that an official 1942 guide to the city described it as "one of the foremost industrial locations of the Reich" and in 1944 the German Army High Command's Weapons Office listed 127 medium-to-large factories and workshops that were supplying the army with materiel. Nonetheless, according to some historians, the contribution of Dresden to the German war effort may not have been as significant as the planners thought.
The US Air Force Historical Division wrote a report in response to the international concern about the bombing that remained classified until December 1978. It said that there were 110 factories and 50,000 workers in the city supporting the German war effort at the time of the raid. According to the report, there were aircraft components factories; a poison gas factory (Chemische Fabrik Goye and Company); an anti-aircraft and field gun factory (Lehman); an optical goods factory (Zeiss Ikon AG); and factories producing electrical and X-ray apparatus ( AG); gears and differentials (Saxoniswerke); and electric gauges (Gebrüder Bassler). It also said there were barracks, hutted camps, and a munitions storage depot.
The USAF report also states that two of Dresden's traffic routes were of military importance: north-south from Germany to Czechoslovakia, and east–west along the central European uplands. The city was at the junction of the Berlin-Prague-Vienna railway line, as well as the Munich-Breslau, and Hamburg-Leipzig lines. Colonel Harold E. Cook, a US POW held in the Friedrichstadt marshaling yard the night before the attacks, later said that "I saw with my own eyes that Dresden was an armed camp: thousands of German troops, tanks and artillery and miles of freight cars loaded with supplies supporting and transporting German logistics towards the east to meet the Russians".
An RAF memo issued to airmen on the night of the attack gave some reasoning for the raid:
In the raid, major industrial areas in the suburbs, which stretched for miles, were not targeted. According to historian Donald Miller, "the economic disruption would have been far greater had Bomber Command targeted the suburban areas where most of Dresden's manufacturing might was concentrated".
The attacks
Night of 13/14 February
The Dresden attack was to have begun with a USAAF Eighth Air Force bombing raid on 13 February 1945. The Eighth Air Force had already bombed the railway yards near the centre of the city twice in daytime raids: once on 7 October 1944 with 70 tons of high-explosive bombs killing more than 400, then again with 133 bombers on 16 January 1945, dropping 279 tons of high-explosives and 41 tons of incendiaries.
On 13 February 1945, bad weather over Europe prevented any USAAF operations, and it was left to RAF Bomber Command to carry out the first raid. It had been decided that the raid would be a double strike, in which a second wave of bombers would attack three hours after the first, just as the rescue teams were trying to put out the fires. As was standard practice, other raids were carried out that night to confuse German air defences. Three hundred and sixty heavy bombers (Lancasters and Halifaxes) bombed a synthetic oil plant in Böhlen, from Dresden, while 71 de Havilland Mosquito medium bombers attacked Magdeburg with small numbers of Mosquitos carrying out nuisance raids on Bonn, Misburg near Hanover and Nuremberg.
When Polish crews of the designated squadrons were preparing for the mission, the terms of the Yalta agreement were made known to them. There was a huge uproar, since the Yalta agreement handed parts of Poland over to the Soviet Union. There was talk of mutiny among the Polish pilots, and their British officers removed their side arms. The Polish Government ordered the pilots to follow their orders and fly their missions over Dresden, which they did.
The first of the British aircraft took off at around 17:20 hours CET for the journey. This was a group of Lancasters from Bomber Command's 83 Squadron, No. 5 Group, acting as the Pathfinders, or flare force, whose job it was to find Dresden and drop magnesium parachute flares, known to the Germans as "Christmas trees", to mark and light up Dresden for the aircraft that would mark the target itself. The next set of aircraft to leave England were twin-engined Mosquito marker planes, which would identify target areas and drop target indicators (TIs) that marked the target for the bombers to aim at. The attack was to centre on the Ostragehege sports stadium, next to the city's medieval Altstadt (old town), with its congested and highly combustible timbered buildings.
The main bomber force, called Plate Rack, took off shortly after the Pathfinders. This group of 254 Lancasters carried 500 tons of high explosives and 375 tons of incendiaries ("fire bombs"). There were 200,000 incendiaries in all, with the high-explosive bombs ranging in weight from —the two-ton "cookies", also known as "blockbusters", because they could destroy an entire large building or street. The high explosives were intended to rupture water mains and blow off roofs, doors, and windows to expose the interiors of the buildings and create an air flow to feed the fires caused by the incendiaries that followed.
The Lancasters crossed into France near the Somme, then into Germany just north of Cologne. At 22:00 hours, the force heading for Böhlen split away from Plate Rack, which turned south east toward the Elbe. By this time, ten of the Lancasters were out of service, leaving 244 to continue to Dresden.
The sirens started sounding in Dresden at 21:51 (CET). the 'Master Bomber' Wing Commander Maurice Smith, flying in a Mosquito, gave the order to the Lancasters: "Controller to Plate Rack Force: Come in and bomb glow of red target indicators as planned. Bomb the glow of red TIs as planned". The first bombs were released at 22:13, the last at 22:28, the Lancasters delivering 881.1 tons of bombs, 57% high explosive, 43% incendiaries. The fan-shaped area that was bombed was long, and at its extreme about wide. The shape and total devastation of the area was created by the bombers of No. 5 Group flying over the head of the fan (Ostragehege stadium) on prearranged compass bearings and releasing their bombs at different prearranged times.
The second attack, three hours later, was by Lancaster aircraft of 1, 3, 6 and 8 Groups, 8 Group being the Pathfinders. By now, the thousands of fires from the burning city could be seen more than away on the ground, and away in the air, with smoke rising to . The Pathfinders therefore decided to expand the target, dropping flares on either side of the firestorm, including the Hauptbahnhof, the main train station, and the Großer Garten, a large park, both of which had escaped damage during the first raid. The German sirens sounded again at 01:05, but as there was practically no electricity, these were small hand-held sirens that were heard within only a block. Between 01:21 and 01:45, 529 Lancasters dropped more than 1,800 tons of bombs.
14–15 February
On the morning of 14 February 431 United States Army Air Force bombers of the Eighth Air Force's 1st Bombardment Division were scheduled to bomb Dresden near midday, and the 457 aircraft of 3rd Bombardment Division were to follow to bomb Chemnitz, while the 375 bombers of the 2nd Bombardment Division would bomb a synthetic oil plant in Magdeburg. Another 84 bombers would attack Wesel. The bomber groups were protected by 784 North American P-51 Mustangs of the Eighth Air Force's VIII Fighter Command, 316 of which covered the Dresden attack - a total of almost 2,100 Eighth Army Air Force aircraft over Saxony during 14 February.
Primary sources disagree as to whether the aiming point was the marshalling yards near the centre of the city or the centre of the built-up urban area. The report by the 1st Bombardment Division's commander to his commander states that the targeting sequence was the centre of the built-up area in Dresden if the weather was clear. If clouds obscured Dresden but Chemnitz was clear, Chemnitz was the target. If both were obscured, they would bomb the centre of Dresden using H2X radar. The mix of bombs for the Dresden raid was about 40% incendiaries—much closer to the RAF city-busting mix than the USAAF usually used in precision bombardment. Taylor compares this 40% mix with the raid on Berlin on 3 February, where the ratio was 10% incendiaries. This was a common mix when the USAAF anticipated cloudy conditions over the target.
316 B-17 Flying Fortresses bombed Dresden, dropping 771 tons of bombs. The remaining 115 bombers from the stream of 431 misidentified their targets. Sixty bombed Prague, dropping 153 tons of bombs, while others bombed Brux and Pilsen. The 379th bombardment group started to bomb Dresden at 12:17, aiming at marshalling yards in the Friedrichstadt district west of the city centre, as the area was not obscured by smoke and cloud. The 303rd group arrived over Dresden two minutes after the 379th and found their view obscured by clouds, so they bombed Dresden using H2X radar. The groups that followed the 303rd (92nd, 306th, 379th, 384th and 457th) also found Dresden obscured by clouds, and they too used H2X. H2X aiming caused the groups to bomb with a wide dispersal over the Dresden area. The last group to attack Dresden was the 306th, and they finished by 12:30.
Strafing of civilians has become a traditional part of the oral history of the raids, since a March 1945 article in the Nazi-run weekly newspaper Das Reich claimed this had occurred. Historian Götz Bergander, an eyewitness to the raids, found no reports on strafing for 13–15 February by any pilots or the German military and police. He asserted in Dresden im Luftkrieg (1977) that only a few tales of civilians being strafed were reliable in detail, and all were related to the daylight attack on 14 February. He concluded that some memory of eyewitnesses was real, but that it had misinterpreted the firing in a dogfight as deliberately aimed at people on the ground. In 2000, historian Helmut Schnatz found an explicit order to RAF pilots not to strafe civilians on the way back from Dresden. He also reconstructed timelines with the result that strafing would have been almost impossible due to lack of time and fuel. Frederick Taylor in Dresden (2004), basing most of his analysis on the work of Bergander and Schnatz, concludes that no strafing took place, although some stray bullets from aerial dogfights may have hit the ground and been mistaken for strafing by those in the vicinity. The official historical commission collected 103 detailed eyewitness accounts and let the local bomb disposal services search according to their assertions. They found no bullets or fragments that would have been used by planes of the Dresden raids.
On 15 February, the 1st Bombardment Division's primary target—the Böhlen synthetic oil plant near Leipzig—was obscured by clouds, so its groups diverted to their secondary target, Dresden. Dresden was also obscured by clouds, so the groups targeted the city using H2X. The first group to arrive over the target was the 401st, but it missed the city centre and bombed Dresden's southeastern suburbs, with bombs also landing on the nearby towns of Meissen and Pirna. The other groups all bombed Dresden between 12:00 and 12:10. They failed to hit the marshalling yards in the Friedrichstadt district and, as in the previous raid, their ordnance was scattered over a wide area.
German defensive action
Dresden's air defences had been depleted by the need for more weaponry to fight the Red Army, and the city lost its last massive flak battery in January 1945. By this point in the war, the Luftwaffe was severely hampered by a shortage of both pilots and aircraft fuel; the German radar system was also degraded, lowering the warning time to prepare for air attacks. The RAF also had an advantage over the Germans in the field of electronic radar countermeasures.
Of 796 British bombers that participated in the raid, six were lost, three of those hit by bombs dropped by aircraft flying over them. On the following day, only a single US bomber was shot down, as the large escort force was able to prevent Luftwaffe day fighters from disrupting the attack.
On the ground
The sirens started sounding in Dresden at 21:51 (CET). Frederick Taylor writes that the Germans could see that a large enemy bomber formation—or what they called "ein dicker Hund" (lit: a fat dog, a "major thing")—was approaching somewhere in the east. At 21:39 the Reich Air Defence Leadership issued an enemy aircraft warning for Dresden, although at that point it was thought Leipzig might be the target. At 21:59 the Local Air Raid Leadership confirmed that the bombers were in the area of Dresden-Pirna. Taylor writes the city was largely undefended; a night fighter force of ten Messerschmitt Bf 110Gs at Klotzsche airfield was scrambled, but it took them half an hour to get into an attack position. At 22:03 the Local Air Raid Leadership issued the first definitive warning: "Warning! Warning! Warning! The lead aircraft of the major enemy bomber forces have changed course and are now approaching the city area". Some 10,000 fled to the great open space of the Grosse Garten, the magnificent royal park of Dresden, nearly one and a half square miles in all. Here they were caught by the second raid, which started without an air-raid warning, at 1:22 a.m. At 11:30 a.m., the third wave of bombers, the two hundred eleven American Flying Fortresses, began their attack.
There were few public air raid shelters. The largest, beneath the main railway station, housed 6,000 refugees. As a result, most people took shelter in cellars, but one of the air raid precautions the city had taken was to remove thick cellar walls between rows of buildings and replace them with thin partitions that could be knocked through in an emergency. The idea was that, as one building collapsed or filled with smoke, those sheltering in the basements could knock walls down and move into adjoining buildings. With the city on fire everywhere, those fleeing from one burning cellar simply ran into another, with the result that thousands of bodies were found piled up in houses at the ends of city blocks.
A Dresden police report written shortly after the attacks reported that the old town and the inner eastern suburbs had been engulfed in a single fire that had destroyed almost 12,000 dwellings. The same report said that the raids had destroyed the Wehrmacht's main command post in the Taschenbergpalais, 63 administration buildings, the railways, 19 military hospitals, 19 ships and barges, and a number of less significant military facilities. Six hundred and forty shops, 64 warehouses, 39 schools, 31 stores, 31 large hotels, 26 public houses/bars, 26 insurance buildings, 24 banks, 19 postal facilities, 19 hospitals and private clinics including auxiliary, overflow hospitals, 18 cinemas, 11 churches and 6 chapels, 5 consulates, 4 tram facilities, 3 theatres, 2 market halls, the zoo, the waterworks, and 5 other cultural buildings were also destroyed. Almost 200 factories were damaged, 136 seriously (including several of the Zeiss Ikon precision optical engineering works), 28 with medium to serious damage, and 35 with light damage.
An RAF assessment showed that 23% of the industrial buildings and 56% of the non-industrial buildings, not counting residential buildings, had been seriously damaged. Around 78,000 dwellings had been completely destroyed; 27,700 were uninhabitable, and 64,500 damaged but readily repairable.
During his post-war interrogation, Albert Speer, Reich Minister of Armaments and War Production, said that Dresden's industrial recovery from the bombings was rapid.
Fatalities
According to the official German report Tagesbefehl (Order of the Day) no. 47 ("TB47") issued on 22 March, the number of dead recovered by that date was 20,204, including 6,865 who were cremated on the Altmarkt square, and they expected the total number of deaths to be about 25,000. Another report on 3 April put the number of corpses recovered at 22,096. Three municipal and 17 rural cemeteries outside Dresden recorded up to 30 April 1945 a total of at least 21,895 buried bodies from the Dresden raids, including those cremated on the Altmarkt.
Between 100,000 and 200,000 refugees fleeing westward from advancing Soviet forces were in the city at the time of the bombing. Exact figures are unknown, but reliable estimates were calculated based on train arrivals, foot traffic, and the extent to which emergency accommodation had to be organised. The city authorities did not distinguish between residents and refugees when establishing casualty numbers and "took great pains to count all the dead, identified and unidentified". This was largely achievable because most of the dead succumbed to suffocation; in only four places were recovered remains so badly burned that it was impossible to ascertain the number of victims. The uncertainty this introduced is thought to amount to no more than 100 people. 35,000 people were registered with the authorities as missing after the raids, around 10,000 of whom were later found alive.
A further 1,858 bodies were discovered during the reconstruction of Dresden between the end of the war and 1966. Since 1989, despite extensive excavation for new buildings, no new war-related bodies have been found. Seeking to establish a definitive casualty figure, in part to address propagandisation of the bombing by far-right groups, the Dresden city council in 2005 authorised an independent Historians' Commission (Historikerkommission) to conduct a new, thorough investigation, collecting and evaluating available sources. The results were published in 2010 and stated that between 22,700 and 25,000 people had been killed.
Wartime political responses
German
Development of a German political response to the raid took several turns. Initially, some of the leadership, especially Robert Ley and Joseph Goebbels, wanted to use it as a pretext for abandonment of the Geneva Conventions on the Western Front. In the end, the only political action the German government took was to exploit it for propaganda purposes. Goebbels is reported to have wept with rage for twenty minutes after he heard the news of the catastrophe, before launching into a bitter attack on Hermann Göring, the commander of the Luftwaffe: "If I had the power I would drag this cowardly good-for-nothing, this Reich marshal, before a court. ... How much guilt does this parasite not bear for all this, which we owe to his indolence and love of his own comforts....".
On 16 February, the Propaganda Ministry issued a press release that claimed that Dresden had no war industries; it was a city of culture.
On 25 February, a new leaflet with photographs of two burned children was released under the title "Dresden—Massacre of Refugees", stating that 200,000 had died. Since no official estimate had been developed, the numbers were speculative, but newspapers such as the Stockholm Svenska Morgonbladet used phrases such as "privately from Berlin," to explain where they had obtained the figures. Frederick Taylor states that "there is good reason to believe that later in March copies of—or extracts from—[an official police report] were leaked to the neutral press by Goebbels's Propaganda Ministry ... doctored with an extra zero to increase [the total dead from the raid] to 202,040". On 4 March, Das Reich, a weekly newspaper founded by Goebbels, published a lengthy article emphasising the suffering and destruction of a cultural icon, without mentioning damage to the German war effort.
Taylor writes that this propaganda was effective, as it not only influenced attitudes in neutral countries at the time, but also reached the House of Commons, when Richard Stokes, a Labour Member of Parliament, and a long term opponent of area-bombing, quoted information from the German Press Agency (controlled by the Propaganda Ministry). It was Stokes's questions in the House of Commons that were in large part responsible for the shift in British opinion against this type of raid. Taylor suggests that, although the destruction of Dresden would have affected people's support for the Allies regardless of German propaganda, at least some of the outrage did depend on Goebbels' falsification of the casualty figures.
British
The destruction of the city provoked unease in intellectual circles in Britain. According to Max Hastings, by February 1945, attacks upon German cities had become largely irrelevant to the outcome of the war and the name of Dresden resonated with cultured people all over Europe—"the home of so much charm and beauty, a refuge for Trollope's heroines, a landmark of the Grand Tour." He writes that the bombing was the first time the public in Allied countries seriously questioned the military actions used to defeat the Germans.
The unease was made worse by an Associated Press story that the Allies had resorted to terror bombing. At a press briefing held by the Supreme Headquarters Allied Expeditionary Force two days after the raids, British Air Commodore Colin McKay Grierson told journalists:
One of the journalists asked whether the principal aim of bombing Dresden would be to cause confusion among the refugees or to blast communications carrying military supplies. Grierson answered that the primary aim was to attack communications to prevent the Germans from moving military supplies, and to stop movement in all directions if possible. He then added in an offhand remark that the raid also helped destroy "what is left of German morale." Howard Cowan, an Associated Press war correspondent, subsequently filed a story claiming that the Allies had resorted to terror bombing. There were follow-up newspaper editorials on the issue and a longtime opponent of strategic bombing, Richard Stokes MP, asked questions in the House of Commons on 6 March.
Churchill subsequently re-evaluated the goals of the bombing campaigns, to focus less on strategic targets, and more toward targets of tactical significance. On 28 March, in a memo sent by telegram to General Ismay for the British Chiefs of Staff and the Chief of the Air Staff, he wrote:
Having been given a paraphrased version of Churchill's memo by Bottomley, on 29 March, Air Chief Marshal Arthur Harris wrote to the Air Ministry:
The phrase "worth the bones of one British grenadier" echoed Otto von Bismarck's: "The whole of the Balkans is not worth the bones of a single Pomeranian grenadier". Under pressure from the Chiefs of Staff and in response to the views expressed by Portal and Harris among others, Churchill withdrew his memo and issued a new one. This was completed on 1 April 1945:
American
John Kenneth Galbraith was among those in the Roosevelt administration who had qualms about the bombing. As one of the directors of the United States Strategic Bombing Survey, formed late in the war by the American Office of Strategic Services to assess the results of the aerial bombardments of Nazi Germany, he wrote: "The incredible cruelty of the attack on Dresden when the war had already been won—and the death of children, women, and civilians—that was extremely weighty and of no avail". The Survey's majority view on the Allies' bombing of German cities, however, concluded:
Timeline
Reconstruction and reconciliation
After the war, and again after German reunification, great efforts were made to rebuild some of Dresden's former landmarks, such as the Frauenkirche, the Semperoper (the Saxony state opera house) and the Zwinger Palace (the latter two were rebuilt before reunification).
In 1956, Dresden entered a twin-town relationship with Coventry. As a centre of military and munitions production, Coventry suffered some of the worst attacks on any British city at the hands of the Luftwaffe during the Coventry Blitzes of 1940 and 1941, which killed over 1,200 civilians and destroyed its cathedral.
In 1990, after the fall of the Berlin Wall, a group of prominent Dresdeners formed an international appeal known as the "Call from Dresden" to request help in rebuilding the Lutheran Frauenkirche, the destruction of which had over the years become a symbol of the bombing. The baroque Church of Our Lady (completed in 1743) had initially appeared to survive the raids, but collapsed a few days later, and the ruins were left in place by later Communist governments as an anti-war memorial.
A British charity, the Dresden Trust, was formed in 1993 to raise funds in response to the call for help, raising £600,000 from 2,000 people and 100 companies and trusts in Britain. One of the gifts they made to the project was an eight-metre high orb and cross made in London by goldsmiths Gant MacDonald, using medieval nails recovered from the ruins of the roof of Coventry Cathedral, and crafted in part by Alan Smith, the son of a pilot who took part in the raid.
The new Frauenkirche was reconstructed over seven years by architects using 3D computer technology to analyse old photographs and every piece of rubble that had been kept and was formally consecrated on 30 October 2005, in a service attended by some 1,800 guests, including Germany's president, Horst Köhler; previous and current chancellors, Gerhard Schröder and Angela Merkel; and the Duke of Kent.
A further development towards the reconstruction of Dresden's historical core came in 1999 when the Dresden Historical Neumarkt Society was founded. The society is committed to reconstructing the historic city centre as much as possible. When plans for the rebuilding of Dresden's Frauenkirche became certain, the (GHND) began calls for the reconstruction of the historic buildings that surrounded it.
In 2003, a petition in support of reconstructing the Neumarkt area was signed by nearly 68,000 people, amounting to 15% of the entire electorate. This had ground breaking results because it demonstrated a broad support for the aims of the initiative and a widespread appreciation for historical Dresden. This led to the city councils decision to rebuild a large amount of baroque buildings in accordance to historical designs, but with modern buildings in between them.
The reconstruction of the surrounding Neumarkt buildings continues to this day.
Post-war debate
The bombing of Dresden remains controversial and is subject to an ongoing debate by historians and scholars regarding the moral and military justifications surrounding the event. British historian Frederick Taylor wrote of the attacks: "The destruction of Dresden has an epically tragic quality to it. It was a wonderfully beautiful city and a symbol of baroque humanism and all that was best in Germany. It also contained all of the worst from Germany during the Nazi period. In that sense it is an absolutely exemplary tragedy for the horrors of 20th century warfare and a symbol of destruction".
Several factors have made the bombing a unique point of contention and debate. First among these are the Nazi government's exaggerated claims immediately afterwards, which drew upon the beauty of the city, its importance as a cultural icon; the deliberate creation of a firestorm; the number of victims; the extent to which it was a necessary military target; and the fact that it was attacked toward the end of the war, raising the question of whether the bombing was needed to hasten the end.
Legal considerations
The Hague Conventions, addressing the codes of wartime conduct on land and at sea, were adopted before the rise of air power. Despite repeated diplomatic attempts to update international humanitarian law to include aerial warfare, it was not updated before the outbreak of World War II. The absence of specific international humanitarian law does not mean that the laws of war did not cover aerial warfare, but the existing laws remained open to interpretation. Specifically, whether the attack can be considered a war crime depends on whether the city was defended and whether resistance was offered against an approaching enemy. Allied arguments centre around the existence of a local air defence system and additional ground defences the Germans were constructing in anticipation of Soviet advances.
Falsification of evidence
The bombing of Dresden has been used by Holocaust deniers and pro-Nazi polemicists—most notably by British writer David Irving—in an attempt to establish a moral equivalence between the war crimes committed by the Nazi government and the killing of German civilians by Allied bombing raids. As such, "grossly inflated" casualty figures have been promulgated over the years, many based on a figure of over 200,000 deaths quoted in a forged version of the casualty report, Tagesbefehl No. 47, that originated with Hitler's Reich Minister of Propaganda Joseph Goebbels.
Marshall inquiry
An inquiry conducted at the behest of U.S. Army Chief of Staff, General George C. Marshall, stated the raid was justified by the available intelligence. The inquiry declared the elimination of the German ability to reinforce a counter-attack against Marshal Konev's extended line or, alternatively, to retreat and regroup using Dresden as a base of operations, were important military objectives. As Dresden had been largely untouched during the war due to its location, it was one of the few remaining functional rail and communications centres. A secondary objective was to disrupt the industrial use of Dresden for munitions manufacture, which American intelligence believed was the case. The shock to military planners and to the Allied civilian populations of the German counterattack known as the Battle of the Bulge had ended speculation that the war was almost over, and may have contributed to the decision to continue with the aerial bombardment of German cities.
The inquiry concluded that by the presence of active German military units nearby, and the presence of fighters and anti-aircraft within an effective range, Dresden qualified as "defended". By this stage in the war both the British and the Germans had integrated air defences at the national level. The German national air-defence system could be used to argue—as the tribunal did—that no German city was "undefended".
Marshall's tribunal declared that no extraordinary decision was made to single out Dresden (for instance, to take advantage of a large number of refugees, or purposely terrorise the German populace). It was argued that the area bombing was intended to disrupt communications and destroy industrial production. The American inquiry established that the Soviets, under allied agreements for the United States and the United Kingdom to provide air support for the Soviet offensive toward Berlin, had requested area bombing of Dresden to prevent a counterattack through Dresden, or the use of Dresden as a regrouping point following a German strategic retreat.
U.S. Air Force Historical Division report
A report by the U.S. Air Force Historical Division (USAFHD) analysed the circumstances of the raid and concluded that it was militarily necessary and justified, based on the following points:
The raid had legitimate military ends, brought about by exigent military circumstances.
Military units and anti-aircraft defences were sufficiently close that it was not valid to consider the city "undefended."
The raid did not use extraordinary means but was comparable to other raids used against comparable targets.
The raid was carried out through the normal chain of command, pursuant to directives and agreements then in force.
The raid achieved the military objective, without excessive loss of civilian life.
The first point regarding the legitimacy of the raid depends on two claims: first, that the railyards subjected to American precision bombing were an important logistical target, and that the city was also an important industrial centre. Even after the main firebombing, there were two further raids on the Dresden railway yards by the USAAF. The first was on 2 March 1945, by 406 B-17s, which dropped 940 tons of high-explosive bombs and 141 tons of incendiaries. The second was on 17 April, when 580 B-17s dropped 1,554 tons of high-explosive bombs and 165 tons of incendiaries.
As far as Dresden being a militarily significant industrial centre, an official 1942 guide described the German city as "...one of the foremost industrial locations of the Reich," and in 1944, the German Army High Command's Weapons Office listed 127 medium-to-large factories and workshops that supplied materiel to the military. Dresden was the seventh largest German city, and by far the largest un-bombed built-up area left, and thus was contributing to the defence of Germany itself.
According to the USAFHD, there were 110 factories and 50,000 workers supporting the German war effort in Dresden at the time of the raid. These factories manufactured fuses and bombsights (at Zeiss Ikon A.G.), aircraft components, anti-aircraft guns, field guns, and small arms, poison gas, gears and differentials, electrical and X-ray apparatus, electric gauges, gas masks, Junkers aircraft engines, and Messerschmitt fighter cockpit parts.
The second of the five points addresses the prohibition in the Hague Conventions, of "attack or bombardment" of "undefended" towns. The USAFHD report states that Dresden was protected by anti-aircraft defences, antiaircraft guns, and searchlights, under the Combined Dresden (Corps Area IV) and Berlin (Corps Area III) Air Service Commands.
The third and fourth points say that the size of the Dresden raid—in terms of numbers, types of bombs and the means of delivery—were commensurate with the military objective and similar to other Allied bombings. On 23 February 1945, the Allies bombed Pforzheim and caused an estimated 20,000 civilian fatalities. The most devastating raid on any city was on Tokyo on 9–10 March (the Meetinghouse raid) which caused over 100,000 casualties, many civilian. The tonnage and types of bombs listed in the service records of the Dresden raid were comparable to (or less than) throw weights of bombs dropped in other air attacks carried out in 1945. In the case of Dresden, as in many other similar attacks, the hour break in between the RAF raids was a deliberate ploy to attack the fire fighters, medical teams, and military units.
In late July 1943, the city of Hamburg was bombed during Operation Gomorrah by combined RAF and USAAF strategic bomber forces. Four major raids were carried out in the span of 10 days, of which the most notable, on the night of 27–28 July, created a devastating firestorm effect similar to Dresden's, killing an estimated 18,474 people. The death toll for that night is included in the overall estimated total of 37,000 for the series of raids. Two-thirds of the remaining population reportedly fled the city after the raids.
The fifth point is that the firebombing achieved the intended effect of disabling the industry in Dresden. It was estimated that at least 23% of the city's industrial buildings were destroyed or severely damaged. The damage to other infrastructure and communications was immense, which would have severely limited the potential use of Dresden to stop the Soviet advance. The report concludes with:
Arguments against justification
Military reasons
The journalist Alexander McKee cast doubt on the meaningfulness of the list of targets mentioned in the 1953 USAF report, pointing out that the military barracks listed as a target were a long way out of the city and were not targeted during the raid. The "hutted camps" mentioned in the report as military targets were also not military but were camps for refugees. It is also stated that the important Autobahn bridge to the west of the city was not targeted or attacked, and that no railway stations were on the British target maps, nor any bridges, such as the railway bridge spanning the Elbe River. Commenting on this, McKee says: "The standard whitewash gambit, both British and American, is to mention that Dresden contained targets X, Y and Z, and to let the innocent reader assume that these targets were attacked, whereas in fact the bombing plan totally omitted them and thus, except for one or two mere accidents, they escaped". McKee further asserts "The bomber commanders were not really interested in any purely military or economic targets, which was just as well, for they knew very little about Dresden; the RAF even lacked proper maps of the city. What they were looking for was a big built-up area which they could burn, and that Dresden possessed in full measure."
According to historian Sönke Neitzel, "it is difficult to find any evidence in German documents that the destruction of Dresden had any consequences worth mentioning on the Eastern Front. The industrial plants of Dresden played no significant role in German industry at this stage in the war". Wing Commander H. R. Allen said, "The final phase of Bomber Command's operations was far and away the worst. Traditional British chivalry and the use of minimum force in war was to become a mockery and the outrages perpetrated by the bombers will be remembered a thousand years hence".
As an immoral act, but not a war crime
Frederick Taylor told Der Spiegel, "I personally find the attack on Dresden horrific. It was overdone, it was excessive and is to be regretted enormously," but, "A war crime is a very specific thing which international lawyers argue about all the time and I would not be prepared to commit myself nor do I see why I should. I'm a historian." Similarly, British philosopher A. C. Grayling has described RAF area bombardment as an "immoral act" and "moral crime" because "destroying everything... contravenes every moral and humanitarian principle debated in connection with the just conduct of war", though Grayling insisted that it "is not strictly correct to describe area bombing as a 'war crime'."
As a war crime
According to Dr. Gregory Stanton, lawyer and president of Genocide Watch:
Historian Donald Bloxham states, "The bombing of Dresden on 13–14 February 1945 was a war crime". He further argues there was a strong prima facie case for trying Winston Churchill among others and a theoretical case Churchill could have been found guilty. "This should be a sobering thought. If, however it is also a startling one, this is probably less the result of widespread understanding of the nuance of international law and more because in the popular mind 'war criminal', like 'paedophile' or 'terrorist', has developed into a moral rather than a legal categorisation".
German author Günter Grass is one of several intellectuals and commentators who have also called the bombing a war crime.
Proponents of this position argue that the devastation from firebombing was greater than anything that could be justified by military necessity alone, and this establishes a prima facie case. The Allies were aware of the effects of firebombing, as British cities had been subject to them during the Blitz. Proponents disagree that Dresden had a military garrison and claim that most of the industry was in the outskirts and not in the targeted city centre, and that the cultural significance of the city should have precluded the Allies from bombing it.
British historian Antony Beevor wrote that Dresden was considered relatively safe, having been spared previous RAF night attacks, and that at the time of the raids there were up to 300,000 refugees in the area seeking sanctuary from the advancing Red Army from the Eastern Front. In Fire Sites, German historian Jörg Friedrich says that the RAF's bombing campaign against German cities in the last months of the war served no military purpose. He claims that Winston Churchill's decision to bomb a shattered Germany between January and May 1945 was a war crime. According to him, 600,000 civilians died during the allied bombing of German cities, including 72,000 children. He claimed some 45,000 people died on one night during the firestorms that engulfed Hamburg in July 1943.
Political response in Germany
Far-right politicians in Germany have sparked a great deal of controversy by promoting the term "Bombenholocaust" ("holocaust by bomb") to describe the raids. Der Spiegel writes that, for decades, the Communist government of East Germany promoted the bombing as an example of "Anglo-American terror," and now the same rhetoric is being used by the far right. An example can be found in the extremist nationalist party Nationaldemokratische Partei Deutschlands (NPD). A party's representative, Jürgen Gansel, described the Dresden raids as "mass murder," and "Dresden's holocaust of bombs". This provoked an outrage in the German parliament and triggered responses from the media. Prosecutors said that it was illegal to call the bombing a holocaust. In 2010, several demonstrations by organisations opposing the far-right blocked a demonstration of far-right organisations.
Phrases like "Bomber-Harris, do it again!", "Bomber-Harris Superstar – Thanks from the red Antifa", and "Deutsche Täter sind keine Opfer!" ("German perpetrators are no victims!") are popular slogans among the so-called "Anti-Germans"—a small radical left-wing political movement in Germany and Austria. In 1995, the fiftieth anniversary of the bombing, Anti-Germans praised the bombing on the grounds that so many of the city's civilians had supported Nazism. Similar rallies take place every year.
In art and popular culture
Kurt Vonnegut
Kurt Vonnegut's novel Slaughterhouse-Five or The Children's Crusade: A Duty-Dance with Death (1969) used some elements from his experiences as a prisoner of war at Dresden during the bombing. The story itself is told through the eyes of Billy Pilgrim, a clear stand-in for Vonnegut himself. His account relates that over 135,000 were killed during the firebombings. Vonnegut recalled "utter destruction" and "carnage unfathomable." The Germans put him and other POWs to work gathering bodies for mass burial. "But there were too many corpses to bury. So instead the Nazis sent in troops with flamethrowers. All these civilians' remains were burned to ashes".
In the special introduction to the 1976 Franklin Library edition of the novel, he wrote:
This experience was also used in several of his other books and is included in his posthumously published stories: Armageddon in Retrospect. The firebombing of Dresden was depicted in George Roy Hill's 1972 movie adaptation of Vonnegut's novel.
The death toll of 135,000 given by Vonnegut was taken from The Destruction of Dresden, a 1963 book by David Irving. In a 1965 letter to The Guardian, Irving later adjusted his estimates even higher, "almost certainly between 100,000 and 250,000", but all these figures were shortly found to be inflated: Irving finally published a correction in The Times in a 1966 letter to the editor lowering it to 25,000, in line with subsequent scholarship. Despite Irving's eventual much lower numbers, and later accusations of generally poor scholarship, the figure popularised by Vonnegut remains in general circulation.
Freeman Dyson, a British (and later American) physicist who had worked as a young man with RAF Bomber Command from July 1943 to the end of the war, wrote in later years:
Dyson later goes on to say: "Since the beginning of the war I had been retreating step by step from one moral position to another, until at the end I had no moral position at all".
Other
The German diarist Victor Klemperer includes a first-hand account of the firestorm in his published works.
The main action of the novel Closely Observed Trains, by Czech author Bohumil Hrabal, takes place on the night of the first raid.
In the 1983 Pink Floyd album The Final Cut, "The Hero's Return", the protagonist lives his years after World War II tormented by "desperate memories", part of him still flying "over Dresden at angels 1–5" (fifteen-thousand feet).
In the song "Tailgunner", Iron Maiden starts with "Trace your way back 50 Years / To the glow of Dresden – blood and tears".
Jonathan Safran Foer's novel Extremely Loud and Incredibly Close (2005) incorporates the bombings into essential parts of the story.
The bombings are a central theme in the 2006 German TV production Dresden by director Roland Suso Richter. Along with the romantic plot between a British bomber pilot and a German nurse, the movie attempts to reconstruct the facts surrounding the Dresden bombings from both the perspective of the RAF pilots and the Germans in Dresden at the time.
The bombing is featured in the 1992 Vincent Ward film, Map of the Human Heart, with the hero, Avik, forced to bale out of his bomber and parachute down into the inferno.
The devastation of Dresden was recorded in the woodcuts of Wilhelm Rudolph, an artist born in the city who resided there until his death in 1982, and was 55 at the time of the bombing. His studio having burned in the attack with his life's work, Rudolph immediately set out to record the destruction, systematically drawing block after block, often repeatedly to show the progress of clearing or chaos that ensued in the ruins. Although the city had been sealed off by the Wehrmacht to prevent looting, Rudolph was granted a special permit to enter and carry out his work, as he would be during the Russian occupation as well. By the end of 1945 he had completed almost 200 drawings, which he transferred to woodcuts following the war. He organised these as discrete series that he would always show as a whole, from the 52 woodcuts of Aus (Out, or Gone) in 1948, the 35 woodcuts Dresden 1945–After the Catastrophe in 1949, and the 15 woodcuts and 5 lithographs of Dresden 1945 in 1955. Of this work, Rudolph later described himself as gripped by an "obsessive-compulsive state," under the preternatural spell of war, which revealed to him that "the utterly fantastic is the reality. ... Beside that, every human invention remains feeble."
In David Alan Mack's The Midnight Front, first book of his secret history historical fantasy series The Dark Arts, the bombing was a concentrated effort by the British, Soviet, and American forces to kill all of the known karcists (sorcerers) in the world in one fell swoop, allied or not, out of fear of their power.
The bombing is featured in the 2018 German film, Never Look Away.
In the movie Airplane II: The Sequel, Sonny Bono's character as the terrorist bomber aboard the space shuttle clutches a briefcase affixed with colorful travel stickers including the Dresden Grand Hotel, Pearl Harbor, Nagasaki, and Hiroshima.
The tragedy of Dresden, as seen through the eyes of Polish forced laborers, was presented by Polish director Jan Rybkowski in the 1961 movie Tonight a City Will Die.
See also
Bombing of Chongqing, 1938–43 – the five years of massive terror-bombings and air battles over the Chinese wartime capital
Bombing of Tokyo (10 March 1945), the firebombing raid on Tokyo codenamed Operation Meetinghouse on 9/10 March 1945
The Blitz – German air raids on British cities in which at least 40,000 died, including 57 consecutive nights of air raids just over London
Baedeker Blitz – Air raids on English cities of cultural/historical importance, rather than military significance
Bombing of Guernica – German/Italian air raid that sparked international outrage
References
Notes
Citations
Bibliography
.
Further reading
External links
Official Memorial
US Air Force Historical Support Division – description, photos
1945 in Germany
1945 in military history
Explosions in 1945
History of Dresden
Firebombings
History of the Royal Air Force during World War II
United States Army Air Forces
Dresden
Dresden
February 1945 events
Germany–United Kingdom military relations
Germany–United States military relations
World War II massacres |
null | null | Triticale | eng_Latn | Triticale (; × Triticosecale) is a hybrid of wheat (Triticum) and rye (Secale) first bred in laboratories during the late 19th century in Scotland and Germany. Commercially available triticale is almost always a second-generation hybrid, i.e., a cross between two kinds of primary (first-cross) triticales. As a rule, triticale combines the yield potential and grain quality of wheat with the disease and environmental tolerance (including soil conditions) of rye. Only recently has it been developed into a commercially viable crop. Depending on the cultivar, triticale can more or less resemble either of its parents. It is grown mostly for forage or fodder, although some triticale-based foods can be purchased at health food stores and can be found in some breakfast cereals.
When crossing wheat and rye, wheat is used as the female parent and rye as the male parent (pollen donor). The resulting hybrid is sterile and must be treated with colchicine to induce polyploidy and thus the ability to reproduce itself.
The primary producers of triticale are Poland, Germany, Belarus, France and Russia. In 2014, according to the Food and Agriculture Organization (FAO), 17.1 million tons were harvested in 37 countries across the world.
The triticale hybrids are all amphidiploid, which means the plant is diploid for two genomes derived from different species. In other words, triticale is an allotetraploid. In earlier years, most work was done on octoploid triticale. Different ploidy levels have been created and evaluated over time. The tetraploids showed little promise, but hexaploid triticale was successful enough to find commercial application.
The International Maize and Wheat Improvement Center (CIMMYT) triticale improvement program was intended to improve food production and nutrition in developing countries. Triticale was thought to have potential in the production of bread and other food products, such as cookies, pasta, pizza dough and breakfast cereals. The protein content is higher than that of wheat, although the glutenin fraction is less. The grain has also been stated to have higher levels of lysine than wheat. Acceptance would require the milling industry to adapt to triticale, as the milling techniques employed for wheat are unsuited to triticale. Past research indicated that triticale could be used as a feed grain and, particularly, later research found that its starch is readily digested. As a feed grain, triticale is already well established and of high economic importance. It has received attention as a potential energy crop, and research is currently being conducted on the use of the crop's biomass in bioethanol production.
History
In the 19th century, crossing cultivars or species became better understood, allowing the controlled hybridization of more plants and animals. Semen could be taken from a male of one animal species to fertilize a female from another species that wouldn't mate with the male directly, and crops plants such as wheat or rye could more systematically be hybridized with each other.
In 1873, Alexander Wilson first managed to manually fertilize the female organs of wheat flowers with rye pollen (male gametes), but found that the resulting plants were sterile, much the way the offspring of a horse and donkey is an infertile mule. Fifteen years later in 1888, a partially-fertile hybrid was produced by "Tritosecale Rimpaui Wittmack". Such hybrids only germinate when the chromosomes spontaneously double.
Unfortunately, "partially fertile" was all that was produced until 1937, when it was discovered that the chemical colchicine, which is used both for general plant germination and on humans with gout, would force chromosome doubling by keeping them from pulling apart during cell division.
Triticale then was viable, though at that point the cost of producing the seeds defeated much of the purpose of producing it.
By the 1960s, triticale was being produced that was far more nutritious than normal wheat. But it was a poorly-producing crop, sometimes yielding shriveled kernels, germinating poorly or prematurely, and didn't bake well.
Modern triticale has overcome most of these problems, after decades of additional breeding and gene transfer with wheat and rye. Millions of acres of the crop are grown around the world, slowly increasing toward becoming a significant source of food-calories.
Species
Triticale hybrids are currently classified by ploidy into three nothospecies:
× Triticosecale semisecale (Mackey) K.Hammer & Filat. – tetraploid triticale. Unstable, but used in breeding bridging. Includes the following crosses:
Triticum monococcum × Secale cereale, genome AARR;
Alternative crosses, genome ABRR (mixogenome A/B).
× Triticosecale neoblaringhemii A.Camus – hexaploid triticale. Stable, currently very successful in argiculture. May be produced by Secale cereale × Triticum turgidum, genome AABBRR.
× Triticosecale rimpaui Wittm. – octaploid triticale. Not completely stable, mainly historical importance. May be produced by Secale cereale × Triticum aestivum, genome AABBDDRR.
The current treatment follows the Mac Key 2005 treatment of Triticum using a broad species concept based on genome composition. Traditional classifications used a narrow species concept based on the treatment of wheats b Dorofeev et al. 1979, and hence produced many more species names. The genome notation follows , with the rye genome notated as R.
Biology and genetics
Earlier work with wheat-rye crosses was difficult due to low survival of the resulting hybrid embryo and spontaneous chromosome doubling. These two factors were difficult to predict and control. To improve the viability of the embryo and thus avoid its abortion, in vitro culture techniques were developed (Laibach, 1925). Colchicine was used as a chemical agent to double the chromosomes. After these developments, a new era of triticale breeding was introduced. Earlier triticale hybrids had four reproductive disorders, namely meiotic instability, high aneuploid frequency, low fertility and shriveled seed (Muntzing 1939; Krolow 1966). Cytogenetical studies were encouraged and well funded to overcome these problems.
It is especially difficult to see the expression of rye genes in the background of wheat cytoplasm and the predominant wheat nuclear genome. This makes it difficult to realise the potential of rye in disease resistance and ecological adaptation. One of the ways to relieve this problem was to produce secalotricum, in which rye cytoplasm was used instead of that from wheat.
Triticale is essentially a self-fertilizing, or naturally inbred, crop. This mode of reproduction results in a more homozygous genome. The crop is, however, adapted to this form of reproduction from an evolutionary point of view. Cross-fertilization is also possible, but it is not the primary form of reproduction.
Sr27 is a stem rust resistance gene which is commonly found in triticale. Originally from rye (Imperial Rye), now () widely found in triticale. Located on the 3A chromosome arm, originally from 3R. Virulence has been observed in field by Puccinia graminis f. sp. secalis (Pgs) and in an artificial cross Pgs Puccinia graminis f. sp. tritici (Pgt). When successful, Sr27 is among the few Srs that does not even allow the underdeveloped uredinia and slight degree of sporulation commonly allowed by most Srs. Instead there are necrotic or chlorotic flecks. Deployment in triticale in New South Wales and Queensland, Australia however rapidly showed virulence between 1982 and 1984 - the first virulence on this gene in the world. (This was especially associated with the cultivar Coorong.) Therefore, CIMMYT's triticale offerings were tested and many were found to depend solely on Sr27. Four years later, in 1988 virulence was found in South Africa. Sr27 has become less common in CIMMYT triticales since the mid-'80s.
Conventional breeding approaches
The aim of a triticale breeding programme is mainly focused on the improvement of quantitative traits, such as grain yield, nutritional quality and plant height, as well as traits which are more difficult to improve, such as earlier maturity and improved test weight (a measure of bulk density). These traits are controlled by more than one gene. Problems arise, however, because such polygenic traits involve the integration of several physiological processes in their expression. Thus the lack of single-gene control (or simple inheritance) results in low trait heritability (Zumelzú et al. 1998).
Since the induction of the CIMMYT triticale breeding programme in 1964, the improvement in realised grain yield has been remarkable. In 1968, at Ciudad Obregón, Sonora, in northwest Mexico, the highest yielding triticale line produced 2.4 t/ha. Today, CIMMYT has released high yielding spring triticale lines (e.g. Pollmer-2) which have surpassed the 10 t/ha yield barrier under optimum production conditions.
Based on the commercial success of other hybrid crops, the use of hybrid triticales as a strategy for enhancing yield in favourable, as well as marginal, environments has proven successful over time. Earlier research conducted by CIMMYT made use of a chemical hybridising agent to evaluate heterosis in hexaploid triticale hybrids. To select the most promising parents for hybrid production, test crosses conducted in various environments are required, because the variance of their specific combining ability under differing environmental conditions is the most important component in evaluating their potential as parents to produce promising hybrids. The prediction of general combining ability of any triticale plant from the performance of its parents is only moderate with respect to grain yield. Commercially exploitable yield advantages of hybrid triticale cultivars is dependent on improving parent heterosis and on advances in inbred-line development.
Triticale is useful as an animal feed grain. However, it is necessary to improve its milling and bread-making quality aspects to increase its potential for human consumption. The relationship between the constituent wheat and rye genomes were noted to produce meiotic irregularities, and genome instability and incompatibility presented numerous problems when attempts were made to improve triticale. This led to two alternative methods to study and improve its reproductive performance, namely, the improvement of the number of grains per floral spikelet and its meiotic behaviour. The number of grains per spikelet has an associated low heritability value (de Zumelzú et al. 1998). In improving yield, indirect selection (the selection of correlated/related traits other than that to be improved) is not necessarily as effective as direct selection. (Gallais 1984)
Lodging (the toppling over of the plant stem, especially under windy conditions) resistance is a complexly inherited (expression is controlled by many genes) trait, and has thus been an important breeding aim in the past. The use of dwarfing genes, known as Rht genes, which have been incorporated from both Triticum and Secale, has resulted in a decrease of up to 20 cm in plant height without causing any adverse effects.
Application of newer techniques
Abundant information exists concerning disease resistance (R) genes for wheat, and a continuously updated on-line catalogue, the Catalogue of Gene Symbols, of these genes can be found at . Another on-line database of cereal rust resistance genes is available at . Unfortunately, less is known about rye and particularly triticale R-genes. Many R-genes have been transferred to wheat from its wild relatives, and appear in the catalogue, thus making them available for triticale breeding. The two mentioned databases are significant contributors to improving the genetic variability of the triticale gene pool through gene (or more specifically, allele) provision. Genetic variability is essential for progress in breeding. In addition, genetic variability can also be achieved by producing new primary triticales, which essentially means the reconstitution of triticale, and the development of various hybrids involving triticale, such as triticale-rye hybrids. In this way, some chromosomes from the R genome have been replaced by some from the D genome. The resulting so-called substitution and translocation triticale facilitates the transfer of R-genes.
Introgression
Introgression involves the crossing of closely related plant relatives, and results in the transfer of 'blocks' of genes, i.e. larger segments of chromosomes compared to single genes. R-genes are generally introduced within such blocks, which are usually incorporated/translocated/introgressed into the distal (extreme) regions of chromosomes of the crop being introgressed. Genes located in the proximal areas of chromosomes may be completely linked (very closely spaced), thus preventing or severely hampering genetic recombination, which is necessary to incorporate such blocks. Molecular markers (small lengths of DNA of a characterized/known sequence) are used to 'tag' and thus track such translocations. A weak colchicine solution has been employed to increase the probability of recombination in the proximal chromosome regions, and thus the introduction of the translocation to that region. The resultant translocation of smaller blocks that indeed carry the R-gene(s) of interest has decreased the probability of introducing unwanted genes.
Production of doubled haploids
Doubled haploid (DH) plants have the potential to save much time in the development of inbred lines. This is achieved in a single generation, as opposed to many, which would otherwise occupy much physical space/facilities. DHs also express deleterious recessive alleles otherwise masked by dominance effects in a genome containing more than one copy of each chromosome (and thus more than one copy of each gene). Various techniques exist to create DHs. The in vitro culture of anthers and microspores is most often used in cereals, including triticale. These two techniques are referred to as androgenesis, which refers to the development of pollen. Many plant species and cultivars within species, including triticale, are recalcitrant in that the success rate of achieving whole newly generated (diploid) plants is very low. Genotype by culture medium interaction is responsible for varying success rates, as is a high degree of microspore abortion during culturing.(Johansson et al. 2000) The response of parental triticale lines to anther culture is known to be correlated to the response of their progeny. Chromosome elimination is another method of producing DHs, and involves hybridisation of wheat with maize (Zea mays L.), followed by auxin treatment and the artificial rescue of the resultant haploid embryos before they naturally abort. This technique is applied rather extensively to wheat. Its success is in large part due to the insensitivity of maize pollen to the crossability inhibitor genes known as Kr1 and Kr2 that are expressed in the floral style of many wheat cultivars. The technique is unfortunately less successful in triticale. However, Imperata cylindrica (a grass) was found to be just as effective as maize with respect to the production of DHs in both wheat and triticale.
Application of molecular markers
An important advantage of biotechnology applied to plant breeding is the speeding up of cultivar release that would otherwise take 8–12 years. It is the process of selection that is actually enhanced, i.e., retaining that which is desirable or promising and ridding that which is not. This carries with it the aim of changing the genetic structure of the plant population. The website is a valuable resource for marker assisted selection (MAS) protocols relating to R-genes in wheat. MAS is a form of indirect selection. The Catalogue of Gene Symbols mentioned earlier is an additional source of molecular and morphological markers. Again, triticale has not been well characterised with respect to molecular markers, although an abundance of rye molecular markers makes it possible to track rye chromosomes and segments thereof within a triticale background.
Yield improvements of up to 20% have been achieved in hybrid triticale cultivars due to heterosis. This raises the question of what inbred lines should be crossed (to produce hybrids) with each other as parents to maximize yield in their hybrid progeny. This is termed the 'combining ability' of the parental lines. The identification of good combining ability at an early stage in the breeding programme can reduce the costs associated with 'carrying' a large number of plants (literally thousands) through it, and thus forms part of efficient selection. Combining ability is assessed by taking into consideration all available information on descent (genetic relatedness), morphology, qualitative (simply inherited) traits and biochemical and molecular markers. Exceptionally little information exists on the use of molecular markers to predict heterosis in triticale. Molecular markers are generally accepted as better predictors than morphological markers (of agronomic traits) due to their insensitivity to variation in environmental conditions.
A useful molecular marker known as a simple sequence repeat (SSR) is used in breeding with respect to selection. SSRs are segments of a genome composed of tandem repeats of a short sequence of nucleotides, usually two to six base pairs. They are popular tools in genetics and breeding because of their relative abundance compared to other marker types, a high degree of polymorphism (number of variants), and easy assaying by polymerase chain reaction. However, they are expensive to identify and develop. Comparative genome mapping has revealed a high degree of similarity in terms of sequence colinearity between closely related crop species. This allows the exchange of such markers within a group of related species, such as wheat, rye and triticale. One study established a 58% and 39% transferability rate to triticale from wheat and rye, respectively. Transferability refers to the phenomenon where the sequence of DNA nucleotides flanking the SSR locus (position on the chromosome) is sufficiently homologous (similar) between genomes of closely related species. Thus, DNA primers (generally, a short sequence of nucleotides used to direct the copying reaction during PCR) designed for one species can be used to detect SSRs in related species. SSR markers are available in wheat and rye, but very few, if any, are available for triticale.
Genetic transformation
The genetic transformation of crops involves the incorporation of 'foreign' genes or, rather, very small DNA fragments compared to introgression discussed earlier. Amongst other uses, transformation is a useful tool to introduce new traits or characteristics into the transformed crop. Two methods are commonly employed: infectious bacterial-mediated (usually Agrobacterium) transfer and biolistics, with the latter being most commonly applied to allopolyploid cereals such as triticale. Agrobacterium-mediated transformation, however, holds several advantages, such as a low level of DNA rearrangement in the transgenic plant, a low number of introduced copies of the transforming DNA, stable integration of an a-priori characterized T-DNA fragment (containing the DNA expressing the trait of interest) and an expected higher level of transgene expression. Triticale has, until recently, only been transformed via biolistics, with a 3.3% success rate. Little has been documented on Agrobacterium-mediated transformation of wheat: while no data existed with respect to triticale until 2005, the success rate in later work was nevertheless low.
Conclusion
Triticale holds much promise as a commercial crop, as it has the potential to address specific problems within the cereal industry. Research of a high standard is currently being conducted worldwide in places like Stellenbosch University in South Africa.
Conventional plant breeding has helped establish triticale as a valuable crop, especially where conditions are less favourable for wheat cultivation. Triticale being a synthesized grain notwithstanding, many initial limitations, such as an inability to reproduce due to infertility and seed shrivelling, low yield and poor nutritional value, have been largely eliminated.
Tissue culture techniques with respect to wheat and triticale have seen continuous improvements, but the isolation and culturing of individual microspores seems to hold the most promise. Many molecular markers can be applied to marker-assisted gene transfer, but the expression of R-genes in the new genetic background of triticale remains to be investigated. More than 750 wheat microsatellite primer pairs are available in public wheat breeding programmes, and could be exploited in the development of SSRs in triticale. Another type of molecular marker, single nucleotide polymorphism (SNP), is likely to have a significant impact on the future of triticale breeding.
Health concerns
Like both its hybrid parents – wheat and rye – triticale contains gluten and is therefore unsuitable for people with gluten-related disorders, such as celiac disease, non-celiac gluten sensitivity and wheat allergy sufferers, among others.
In fiction
An episode of the popular TV series Star Trek "The Trouble with Tribbles" revolved around the protection of a grain developed from triticale. This grain was named "quadro-triticale" by writer David Gerrold at the suggestion of producer Gene Coon, with four distinct lobes per kernel. In that episode Mr. Spock correctly attributes the ancestry of the nonfictional grain to 20th-century Canada.
Indeed, in 1953 the University of Manitoba began the first North American triticale breeding program. Early breeding efforts concentrated on developing a high-yield, drought-tolerant human food crop species suitable for marginal wheat-producing areas. (Later in the episode, Chekov claims that the fictional quadro-triticale was a "Russian invention.")
A later episode titled "More Tribbles, More Troubles", in the animated series, also written by Gerrold, dealt with "quinto-triticale", an improvement on the original, having apparently five lobes per kernel.
Three decades later the spinoff series Deep Space Nine revisited quadro-triticale and the depredations of the Tribbles in the episode Trials and Tribble-ations.
References
External links
Cereals
Pooideae
Energy crops |
null | null | Infant formula | eng_Latn | Infant formula, baby formula or just formula (American English) or baby milk, infant milk ,false milk, or first milk (British English), is a manufactured food designed and marketed for feeding to babies and infants under 12 months of age, usually prepared for bottle-feeding or cup-feeding from powder (mixed with water) or liquid (with or without additional water). The U.S. Federal Food, Drug, and Cosmetic Act (FFDCA) defines infant formula as "a food which purports to be or is represented for special dietary use solely as a food for infants by reason of its simulation of human milk or its suitability as a complete or partial substitute for human milk".
Manufacturers state that the composition of infant formula is designed to be roughly based on a human mother's milk at approximately one to three months postpartum; however, there are significant differences in the nutrient content of these products. The most commonly used infant formulas contain purified cow's milk whey and casein as a protein source, a blend of vegetable oils as a fat source, lactose as a carbohydrate source, a vitamin-mineral mix, and other ingredients depending on the manufacturer. In addition, there are infant formulas using soybean as a protein source in place of cow's milk (mostly in the United States and Great Britain) and formulas using protein hydrolysed into its component amino acids for infants who are allergic to other proteins. An upswing in breastfeeding in many countries has been accompanied by a deferment in the average age of introduction of baby foods (including cow's milk), resulting in both increased breastfeeding and increased use of infant formula between the ages of 3- and 12-months.
A 2001 World Health Organization (WHO) report found that infant formula prepared in accordance with applicable Codex Alimentarius standards was a safe complementary food and a suitable breast milk substitute. In 2003, the WHO and UNICEF published their Global Strategy for Infant and Young Child Feeding, which restated that "processed-food products for...young children should, when sold or otherwise distributed, meet applicable standards recommended by the Codex Alimentarius Commission", and also warned that "lack of breastfeeding—and especially lack of exclusive breastfeeding during the first half-year of life—are important risk factors for infant and childhood morbidity and mortality".
In particular, the use of infant formula in less economically developed countries is linked to poorer health outcomes because of the prevalence of unsanitary preparation conditions, including lack of clean water and lack of sanitizing equipment. A formula-fed child living in unclean conditions is between 6 and 25 times more likely to die of diarrhea and four times more likely to die of pneumonia than a breastfed child. Rarely, use of powdered infant formula (PIF) has been associated with serious illness, and even death, due to infection with Cronobacter sakazakii and other microorganisms that can be introduced to PIF during its production. Although C. sakazakii can cause illness in all age groups, infants are believed to be at greatest risk of infection. Between 1958 and 2006, there have been several dozen reported cases of E. sakazakii infection worldwide. The WHO believes that such infections are under-reported.
Uses, risks and controversies
The use and marketing of infant formula has come under scrutiny. Breastfeeding, including exclusive breastfeeding for the first 6 months of life, is widely advocated as "ideal" for babies and infants, both by health authorities — and accordingly in ethical advertising of infant formula manufacturers.
Despite the recommendation that babies be exclusively breastfed for the first 6 months, less than 40% of infants below this age are exclusively breastfed worldwide. The overwhelming majority of American babies are not exclusively breastfed for this period – in 2005 under 12% of babies were breastfed exclusively for the first 6 months, with over 60% of babies of 2 months of age being fed formula, and approximately one in four breastfed infants having infant formula feeding within two days of birth.
Some studies have shown that use of formula can vary according to the parents' socio-economic status, ethnicity or other characteristics. For example, according to a research conducted in Vancouver, Canada, 82.9% of mothers breastfeed their babies at birth, but the number differed between Caucasians (91.6%) and non-Caucasians (56.8%), with the difference essentially attributed to marital status, education and family income. In the United States, mothers of lower socio-economic status have been found less likely to breastfeed, although this may be partly related to adverse effects of government nutrition supplementation programs that provide subsidies for infant formula.
The use of hydrolysed cow milk baby formula versus standard milk baby formula does not appear to change the risk of allergies or autoimmune diseases.
Use of infant formula
In some cases, breastfeeding is medically contraindicated. These include:
Mother's health: The mother is infected with HIV or has active tuberculosis. She is extremely ill or has had certain kinds of breast surgery, which may have removed or disconnected all milk-producing parts of the breast. She is taking any kind of drug that could harm the baby, including both prescription drugs such as cytotoxic chemotherapy for cancer treatments as well as illicit drugs.
One of the main global risks posed by breast milk specifically is the transmission of HIV and other infectious diseases. Breastfeeding by an HIV-infected mother poses a 5–20% chance of transmitting HIV to the baby. However, if a mother has HIV, she is more likely to transmit it to her child during the pregnancy or birth than during breastfeeding. A 2012 study conducted by researchers from the University of North Carolina School of Medicine showed reduced HIV-1 transmission in humanized mice, due to components in the breast milk. Cytomegalovirus infection poses potentially dangerous consequences for pre-term babies. Other risks include mother's infection with HTLV-1 or HTLV-2 (viruses that could cause T-cell leukemia in the baby), herpes simplex when lesions are present on the breasts, and chickenpox in the newborn when the disease manifested in the mother within a few days of birth. In some cases these risks can be mitigated by using heat-treated milk and nursing for a briefer time (e.g. 6 months, rather than 18–24 months), and can be avoided by using an uninfected woman's milk, as via a wet-nurse or milk bank, or by using infant formula and/or treated milk.
In balancing the risks, such as cases where the mother is infected with HIV, a decision to use infant formula versus exclusive breastfeeding may be made based on alternatives that satisfy the “AFASS” (Acceptable, Feasible, Affordable, Sustainable and Safe) principles.
Baby is unable to breastfeed: The child has a birth defect or inborn error of metabolism such as galactosemia that makes breastfeeding difficult or impossible.
Baby is considered at risk for malnutrition: In certain circumstances infants may be at risk for malnutrition, such as due to iron deficiency, vitamin deficiencies (e.g. vitamin D which may be less present in breast milk than needed at high latitudes where there is less sun exposure), or inadequate nutrition during transition to solid foods. Risks can often be mitigated with improved diet and education of mothers and caregivers, including availability of macro and micronutrients. For example, in Canada, marketed infant formulas are fortified with vitamin D, but Health Canada also recommends breastfed infants receive extra vitamin D in the form of a supplement.
Personal preferences, beliefs, and experiences: The mother may dislike breast-feeding or think it is inconvenient. In addition, breastfeeding can be difficult for victims of rape or sexual abuse; for example, it may be a trigger for posttraumatic stress disorder. Many families bottle feed to increase the father's role in parenting his child.
Mental Health: The pressure to breastfeed in many cultures can be so much that the mother’s mental health may take a sharp decline. This can have physical effects such as poor latching as well as milk depletion and a lack of connection to the child. In some cases it is better for the child to be formula fed so that a better bond can be made between mother and child rather than the ‘special bond’ that comes from breastfeeding being tainted by negative breastfeeding experiences. The pressure to breastfeed in many cultures can increase the likelihood of postpartum depression
Absence of the mother: The child is adopted, orphaned, abandoned, or in the sole custody of a man or male same-sex couple. The mother is separated from her child by being in prison or a mental hospital. The mother has left the child in the care of another person for an extended period of time, such as while traveling or working abroad.
Food allergies: The mother eats foods that may provoke an allergic reaction in the infant.
Financial pressures: Maternity leave is unpaid, insufficient, or lacking. The mother's employment interferes with breastfeeding. Mothers who breastfeed may experience a loss of earning power.
Societal structure: Breastfeeding may be forbidden at the mother's job, school, place of worship or in other public places, or the mother may feel that breastfeeding in these places or around other people is immodest, unsanitary, or inappropriate.
Social pressures: Family members, such as mother's husband or boyfriend, or friends or other members of society may encourage the use of infant formula. For example, they may believe that breastfeeding will decrease the mother's energy, health, or attractiveness. Conversely, societal pressures to breastfeed can also lead to mental health issues. A sense of shame from not being able to or struggling to do so equalling being a failure has a connection to Postpartum Depression The widely used phrase “Breast is Best” makes those who are struggling feel extra pressure and shame for not giving their child “the best”. The phrase “Fed is Best” is starting to be used in replacement to relieve some of that pressure and not let women who may be struggling fall into the trap of feeling like bad mothers.
Lack of training and education: The mother lacks education and training from medical providers or community members.
Lactation insufficiency: The mother is unable to produce sufficient milk. In studies that do not account for lactation failure with obvious causes (such as use of formula and/or breast pumps), this affects around 2 to 5% of women. Alternatively, despite a healthy supply, the woman or her family may incorrectly believe that her breast milk is of low quality or in low supply. These women may choose infant formula either exclusively or as a supplement to breastfeeding.
Fear of exposure to environmental contaminants: Certain environmental pollutants, such as polychlorinated biphenyls, can bioaccumulate in the food chain and may be found in humans including mothers' breast milk.
However studies have shown that the greatest risk period for adverse effects from environmental exposures is prenatally. Other studies have further found that the levels of most persistent organohalogen compounds in human milk decreased significantly over the past three decades and equally did their exposure through breastfeeding.
Research on risks from chemical pollution is generally inconclusive in terms of outweighing the benefits of breastfeeding. Studies supported by the WHO and others have found that neurological benefits of breast milk remain, regardless of dioxin exposure.
In developing countries, environmental contaminants associated with increased health risks from use of infant formula, particularly diarrhea due to unclean water and lack of sterile conditions – both prerequisites to the safe use of formula – often outweigh any risks from breastfeeding.
Lack of other sources of breast milk:
Lack of wet nurses: Wet nursing is illegal and stigmatized in some countries, and may not be available. It may also be socially unsupported, expensive, or health screening of wet nurses may not be available. The mother, her doctor, or family may not know that wet nursing is possible, or may believe that nursing by a relative or paid wet-nurse is unhygienic.
Lack of milk banks: Human-milk banks may not be available, as few exist, and many countries cannot provide the necessary screening for diseases and refrigeration.
Health risks
Use of infant formula has been cited for numerous increased health risks. Studies have found infants in developed countries who consume formula are at increased risk for acute otitis media, gastroenteritis, severe lower respiratory tract infections, atopic dermatitis, asthma, obesity, type 1 and 2 diabetes, sudden infant death syndrome (SIDS), eczema and necrotizing enterocolitis when compared to infants who are breastfed. Some studies have found an association between infant formula and lower cognitive development, including iron supplementation in baby formula being linked to lowered I.Q. and other neurodevelopmental delays; however other studies have found no correlation.
Melamine contamination
In 2008, a case of melamine poisoning of infant formula was discovered in China, where milk was deliberately adulterated with the chemical, leading to the death of six babies, and illnesses in more than 300,000 infants, including cases of acute kidney failure. Large quantities of melamine were added to watered-down milk to give it the appearance of having adequate protein levels. Some of those responsible for the poisoning were sentenced to death.
In November 2008, traces of melamine were reported to have been found by the U.S. Food and Drug Administration in infant formula sold in the United States made by the three main American firms — Abbott Laboratories, Nestlé and Mead Johnson — responsible for 90–99% of the infant formula market in that country. The levels were much less than those reported in China, where levels of melamine contamination had reached as much as 2,500 parts per million, about 10,000 times higher than the recorded US levels. The safety data sheet for melamine (CAS registry number 108-78-1; C3-H6-N6) recorded the acute oral toxicity (median lethal dose) at 3161 mg/kg for a rat.
Health Canada conducted a separate test and also detected traces of melamine in infant formula available in Canada. The melamine levels were well below Health Canada's safety limits, although concerns remain about the safety of manufactured food for infants and monitoring of potentially dangerous substances.
Other health controversies
In 1985, Syntex Corporation was ordered to pay $27 million in compensation for the deaths of two American infants who suffered brain damage after drinking the company's baby formula, called Neo-mull-soy. Formulas produced by Syntex had previously been subject to a major recall as they were found to have insufficient chloride to support normal infant growth and development.
In 2003, baby plant-based formula manufactured by the German company Humana and sold in Israel under the brand Remedia caused severe vitamin deficiencies in babies. Babies who consumed the formula were hospitalized with cardiac and neurological symptoms. Three of them died, and at least twenty others were left with severe disabilities. An investigation revealed that the formula contained a much lower quantity of Thiamine than is needed for healthy infant development because of a manufacturing error. Humana's chief food technologist received a 30-month prison sentence for negligent manslaughter in February 2013 over the case.
In 2010, Abbott Laboratories issued a voluntary recall of about five million Similac brand powder infant formulas that were sold in the United States, Guam, Puerto Rico and some Caribbean countries. The recall was issued after the presence of a 'small common beetle' was detected in the product.
In Canada, New Zealand and elsewhere, public concerns have been raised over the continued sale and marketing of soy-based formulae potentially containing high levels of phytoestrogens, linked to abnormal child development including damage to babies' thyroid glands.
In December 2011 Wal-Mart recalled a quantity of infant formula after a baby died in Missouri. "We extend our deepest condolences to this baby boy's family as they try to come to grips with their loss," said Dianna Gee, a Wal-Mart spokeswoman. "As soon as we heard what happened, we immediately reached out to the manufacturer of the formula and to the Department of Health and Senior Services to provide any information we may have to help with the investigation." Wal-Mart said it pulled a batch of Enfamil from its stores nationwide that matched the size and lot number ZP1k7G of the formula that may have sickened the baby in Missouri, Gee said. The baby formula was purchased from a Wal-Mart in Lebanon, Missouri. After the purchase, a 10-day-old infant died from a rare bacterial infection, CNN affiliate KYTV reported. Authorities ran tests to determine if the death came from the formula, the water to make the formula or any other factor, said Mead Johnson Nutrition, the company that makes Enfamil. "We are highly confident in the safety and quality of our products – and the rigorous testing we put them through," said Chris Perille, a Mead Johnson Nutrition spokesman.[Source CNN]
Preparation and content
Variations
Infant formulas come in powder, liquid concentrate, and ready-to-feed forms. They are designed to be prepared by the parent or caregiver in small batches and fed to the infant, usually with either a cup or a baby bottle.
Infant formulas come in a variety of types:
Cow's milk formula is the most commonly used type. The milk has been altered to resemble breast milk.
Soy protein based formulas are frequently used for infants allergic to cow's milk or lactose. Soy-based formulas can also be useful if the parent wants to exclude animal proteins from the child's diet.
Protein hydrolysate formulas contain protein that's been broken down into smaller sizes than are those in cow's milk and soy-based formulas. Protein hydrolysate formulas are meant for babies who do not tolerate cow's milk or soy-based formulas.
Specialized formulas are also available for premature infants and those with specific medical conditions.
Manufacturers and health officials advise it is very important to measure powders or concentrates accurately to achieve the intended final product concentration; otherwise, the child will be malnourished. It is advisable that all equipment that comes into contact with the infant formula be cleaned and sterilized before each use. Proper refrigeration is essential for any infant formula which is prepared in advance.
In developing countries, formula is frequently prepared improperly, resulting in high infant mortality due to malnutrition and diseases such as diarrhea and pneumonia. This is due to lack of clean water, lack of sterile conditions, lack of refrigeration, illiteracy (so written instructions cannot be followed), poverty (diluting formula so that it lasts longer), and lack of education of mothers by formula distributors. These problems and resulting disease and death are a key factor in opposition to the marketing and distribution of infant formula in developing countries by numerous public health agencies and NGOs (discussed in more detail at Nestlé boycott and International Code of Marketing of Breast-milk Substitutes).
Nutritional content
Besides breast milk, infant formula is the only other milk product which the medical community considers nutritionally acceptable for infants under the age of one year (as opposed to cow's milk, goat's milk, or follow-on formula). Supplementing with solid food in addition to breast milk or formula begins during weaning, and most babies begin supplementing about the time their first teeth appear, usually around the age of six months.
Although cow's milk is the basis of almost all infant formula, plain cow's milk is unsuited for infants because of its high casein content and low whey content, and untreated cow's milk is not recommended before the age of 12 months. The infant intestine is not properly equipped to digest non-human milk, and this may often result in diarrhea, intestinal bleeding and malnutrition. To reduce the negative effect on the infant's digestive system, cow's milk used for formula undergoes processing to be made into infant formula. This includes steps to make protein more easily digestible and alter the whey-to-casein protein balance to one closer to human milk, the addition of several essential ingredients (often called "fortification", see below), the partial or total replacement of dairy fat with fats of vegetable or marine origin, etc.
The nutrient content of infant formula for sale in the United States is regulated by the Food and Drug Administration (FDA) based on recommendations by the American Academy of Pediatrics Committee on Nutrition. The following must be included in all formulas produced in the U.S.:
Protein
Fat
Linoleic acid
Vitamins: A, C, D, E, K, thiamin (B1), riboflavin (B2), B6, B12
Niacin
Folic acid
Pantothenic acid
Calcium
Minerals: magnesium, iron, zinc, manganese, copper
Phosphorus
Iodine
Sodium chloride
Potassium chloride
Carbohydrates
Carbohydrates are an important source of energy for growing infants, as they account for 35 to 42% of their daily energy intake. In most cow's milk-based formulas, lactose is the main source of carbohydrates present, but lactose is not present in cow's milk-based lactose-free formulas nor specialized non-milk protein formulas or hydrolyzed protein formulas for infants with milk protein sensitivity. Lactose is also not present in soy-based formulas. Therefore, those formulas without lactose will use other sources of carbohydrates, such as sucrose and glucose, dextrins, and natural and modified starches. Lactose is not only a good source of energy, it also aids in the absorption of the minerals magnesium, calcium, zinc and iron.
Human milk oligosaccharides (HMOs)
HMOs are naturally occurring sugars found in human breast milk, they improve the immune system and act as nutrients to beneficial gut bacteria. Some manufacturers also use human milk oligosaccharides as a modern infant formula supplement to give additional health benefits to their products, however they are not found all types of formula.
Nucleotides
Nucleotides are compounds found naturally in human breast milk. They are involved in critical metabolic processes, such as energy metabolism and enzymatic reactions. Also, as the building blocks of deoxyribonucleic acid (DNA) and ribonucleic acid (RNA), they are essential for normal body functions. Compared to human breast milk, cow's milk has lower levels of the nucleotides uridine, inosine, and cytidine. Therefore, several companies that produce infant formula have added nucleotides to their infant formulas.
Other commonly used ingredients:
Emulsifiers and stabilizers: Ingredients added to prevent the separation of the oil from the water (and its soluble components) in the infant formula. Some commonly used emulsifiers include monoglycerides, diglycerides, and gums.
Diluents: Skim milk is commonly used as the primary diluent in milk-based liquid formula to provide the bulk of the volume. In contrast, purified water is the most commonly used diluent in milk-free formulations.
Policy, industry and marketing
The policy, regulatory and industry environments surrounding the infant formula market vary tremendously between countries.
International
The International Code of Marketing of Breast-milk Substitutes is an international health policy framework adopted by the World Health Assembly of the WHO in 1981 regarding infant formula marketing, including strict restrictions on advertising. Its implementation depends on the laws of different countries and the behavior of infant formula manufacturers – the code has no power itself. Legislation and corporate behavior vary significantly between countries: at least 84 countries have enacted national legislation implementing all or many of the provisions of the Code and 14 countries have draft laws awaiting adoption; whereas elsewhere neither the Code nor its principles are followed by governments or formula manufacturers.
Practices that are banned in the Code include most advertising, claiming health benefits for formula, and giving free samples to women able to breastfeed – this latter practice is particularly criticized because it can interfere with lactation, creating dependence on formula. In many countries free samples of infant formula have been provided to hospitals for decades; infant formula is often the only product routinely provided free of charge to hospitals. The Baby Friendly Hospital Initiative aims to reduce and eliminate this controversial practice.
By country
Philippines
Infant formula is one of the top three consumer commodities in the Philippines, and among the most imported products. Annual sales amount to some US$469 million annually. US$88 million is spent on advertising the product.
Infant formula marketing has been regulated since the 1987 Executive Order 51 or "Milk Code", which regulated, but did not ban, practices such as advertising and providing free samples. Shortly after it was enacted, Wyeth introduced "follow-on formula", which was not in the purview of the Milk Code which predated its market entry.
In 2006, the Department of Health banned the advertising of infant formula and the practice of providing free samples, regardless of intended age group (in the Revised Implementing Rules and Regulations of Executive Order 51, or RIRR). The new regulation was challenged by the infant formula industry in the Supreme Court. Initially the challenge was dismissed, but this decision was reversed following industry pressure and a controversial letter by American business leader Thomas Donahue, then President and CEO of the US Chamber of Commerce, resulting in the regulation being suspended and advertising continuing.
The Guardian newspaper reports widespread illegal advertising and marketing of formula milk contrary to World Health Organization guidelines. Doctors and midwives are encouraged to promote feeding babies formula milk, advertising also targets mothers directly. Babies get sick and sometimes die because poor mothers cannot sterilize bottles.
South Africa
In South Africa, there is a move towards plain packaging of infant formula under R 991 of the Foodstuffs, Cosmetics and Disinfectants Act; as of 6 December 2013, Regulation 7 (Sale and Promotion) is force, whereas Regulations 2-6 (primarily with respect to labelling) are scheduled to come into force on 6 December 2014.
One of the key requirements as per Regulation 3.1.A.iii is a conspicuous message stating “[t]his product shall only be used on the advice of a health professional”.
United Kingdom
In the United Kingdom, infant formula advertising has been allowed since 1995; advertising for "follow-on formula" is legal, which has been cited as a loophole allowing advertising of similarly packaged formula.
United States
In the United States, infant formula is both heavily marketed – the country has not adopted the Code, nor is it being systematically implemented by manufacturers for domestic marketing – and even heavily subsidized by the government: at least one third of the American market is supported by the government, with over half of infant formula sold in the country provided through the Special Supplemental Nutrition Program for Women, Infants, and Children (known as WIC).
According to surveys, over 70% of large U.S. hospitals dispense infant formula to all infants, a practice opposed by the American Academy of Pediatrics and in violation of the Code. The Gerber Products Company began marketing its brand of infant formula directly to the public in October 1989, while the Carnation Company began marketing Good Start infant formula directly to the public in January 1991.
Infant formula costs are a significant fraction of the WIC program costs: 21% post-rebate and 46% pre-rebate. Formula manufacturers are granted a WIC monopoly in individual states. Meanwhile, breastfeeding rates are substantially lower for WIC recipients; this is partly attributed to formula being free of charge to mothers in the WIC program, who are of lower socio-economic status. Violations of federal policy have also been found in terms of infant formula company advertising using the WIC trademark, to reach both WIC and non-WIC participants. In recent years WIC has been expanding its breastfeeding promotion strategies, including providing subsidies for clients who use milk banks.
History
The Wabanaki and other Native American tribal nations of North America did make an infant formula from nuts and cornmeal. Elizabeth Hanson was captured by Wabanaki in 1725 and a Native American woman showed Hanson how to make this infant formula and this she did include in her captivity narrative.
Early infant foods
In 1865, the first infant food was invented.
Throughout history, mothers who could not breastfeed their babies either employed a wet nurse or, less frequently, prepared food for their babies, a process known as "dry nursing". Baby food composition varied according to region and economic status. In Europe and North America during the early 19th century, the prevalence of wet nursing began to decrease, while the practice of feeding babies mixtures based on animal milk rose in popularity. This trend was driven by cultural changes as well as increased sanitation measures, and it continued throughout the 19th and much of the 20th century, with a notable increase after Elijah Pratt invented and patented the India-rubber nipple in 1845. As early as 1846, scientists and nutritionists noted an increase in medical problems and infant mortality was associated with dry nursing. In an attempt to improve the quality of manufactured baby foods, in 1867, Justus von Liebig developed the world's first commercial infant formula, Liebig's Soluble Food for Babies. The success of this product quickly gave rise to competitors such as Mellin's Food, Ridge's Food for Infants and Nestlé's Milk.
Raw milk formulas
As physicians became increasingly concerned about the quality of such foods, medical recommendations such as Thomas Morgan Rotch's "percentage method" (published in 1890) began to be distributed, and gained widespread popularity by 1907. These complex formulas recommended that parents mix cow's milk, water, cream, and sugar or honey in specific ratios to achieve the nutritional balance believed to approximate human milk reformulated in such a way as to accommodate the believed digestive capability of the infant.
At the dawn of the 20th century in the United States, most infants were breastfed, although many received some formula feeding as well. Home-made "percentage method" formulas were more commonly used than commercial formulas in both Europe and the United States. They were less expensive and were widely believed to be healthier. However, formula-fed babies exhibited more diet-associated medical problems, such as scurvy, rickets and bacterial infections than breastfed babies. By 1920, the incidence of scurvy and rickets in formula-fed babies had greatly decreased through the addition of orange juice and cod liver oil to home-made formulas. Bacterial infections associated with formula remained a problem more prevalent in the United States than in Europe, where milk was usually boiled prior to use in formulas.
Evaporated milk formulas
In the 1920s and 1930s, evaporated milk began to be widely commercially available at low prices, and several clinical studies in the period suggested that babies fed evaporated milk formula thrived as well as breastfed babies.
These studies, accompanied by the affordable price of evaporated milk and the availability of the home icebox initiated a tremendous rise in the use of evaporated milk formulas. By the late 1930s, the use of evaporated milk formulas in the United States surpassed all commercial formulas, and by 1950 over half of all babies in the United States were reared on such formulas.
Commercial formulas
In parallel with the enormous shift (in industrialized nations) away from breastfeeding to home-made formulas, nutrition scientists continued to analyze human milk and attempted to make infant formulas that more closely matched its composition. Maltose and dextrins were believed nutritionally important, and in 1912, the Mead Johnson Company released a milk additive called Dextri-Maltose. This formula was made available to mothers only by physicians. In 1919, milkfats were replaced with a blend of animal and vegetable fats as part of the continued drive to closer simulate human milk. This formula was called SMA for "simulated milk adapted."
In the late 1920s, Alfred Bosworth released Similac (for "similar to lactation"), and Mead Johnson released Sobee. Several other formulas were released over the next few decades, but commercial formulas did not begin to seriously compete with evaporated milk formulas until the 1950s. The reformulation and concentration of Similac in 1951, and the introduction (by Mead Johnson) of Enfamil (for "infant milk") in 1959 were accompanied by marketing campaigns that provided inexpensive formula to hospitals and pediatricians. By the early 1960s, commercial formulas were more commonly used than evaporated milk formulas in the United States, which all but vanished in the 1970s. By the early 1970s, over 75% of American babies were fed on formulas, almost entirely commercially produced.
When birth rates in industrial nations tapered off during the 1960s, infant formula companies heightened marketing campaigns in non-industrialized countries. Unfortunately, poor sanitation led to steeply increased mortality rates among infants fed formula prepared with contaminated water. Additionally, a WHO has cited over-diluting formula preparations as resulting in infant malnourishment. Organized protests, the most famous of which was the Nestlé boycott of 1977, called for an end to unethical marketing. This boycott is ongoing, as the current coordinators maintain that Nestlé engages in marketing practices which violate the International Code of Marketing of Breast-milk Substitutes.
Generic brand formulas
In addition to commercially marketed brands, generic brands (or store brands) of infant formula were introduced in the United States in 1997, first by PBM Products. These private label formulas are sold by many leading food and drug retailers such as Wal-Mart, Target, Kroger, Loblaws, and Walgreens. All infant formula brands in the United States are required to adhere to the Food and Drug Administration (FDA) guidelines. As reported by the Mayo Clinic: “as with most consumer products, brand-name infant formulas cost more than generic brands. But that doesn't mean that brand-name [Similac, Nestle, Enfamil] formulas are better. Although manufacturers may vary somewhat in their formula recipes, the FDA requires that all formulas contain the same nutrient density.”
Similarly, in Canada all infant formulas regardless of brand are required to meet standards set by Health Canada.
Follow-on and toddler formulas
Follow-on or toddler formulas are sold for ages 6 months to 3 years (when infants are typically breastfed). In the US, a transition formula is marketed for children from age 9 to 24 months, and a toddler milk is sold for children age 12 to 26 months. In both cases, the ingredients are powdered milk, corn syrup and other added sugars, vegetable oil, and salt.
Toddler formulas are not nutritionally complete, nor are they subject to the same regulations or food labeling laws as infant formula. Critics have argued that follow-on and toddler formulas were introduced to circumvent the regulations regarding infant formula and have resulted in confusing advertising.
An early example of follow-on formula was introduced by Wyeth in the Philippines in 1987, following the introduction in this country of regulations on infant formula advertising, but which did not address follow-on formulas (products that did not exist at the time of their drafting). Similarly, while infant formula advertising is illegal in the United Kingdom, follow-on formula advertising is legal, and the similar packaging and market results in follow-on advertisements frequently being interpreted as advertisements for formula. (See also industry and marketing, below.)
These products have also recently fallen under criticism for contributing to the childhood obesity epidemic in some developed countries due to their marketing and flavoring practices. The drinks are also expensive. Although usually not quite as expensive as infant formula, they can cost four times the price of cow's milk.
Usage since 1970s
Since the early 1970s, industrial countries have witnessed a resurgence in breastfeeding among newborns and infants to 6 months of age. This upswing in breastfeeding has been accompanied by a deferment in the average age of introduction of other foods (such as cow's milk), resulting in increased use of both breastfeeding and infant formula between the ages of 3–12 months.
The global infant formula market has been estimated at $7.9 billion, with North America and Western Europe accounting for 33% of the market and considered largely saturated, and Asia representing 53% of the market. South East Asia is a particularly large fraction of the world market relative to its population. Infant formula is the largest segment of the baby food market, with the fraction given as between 40% and 70%.
Leading health organizations (e.g. WHO, U.S. Centers for Disease Control and Department of Health and Human Services) are attempting to reduce the use of infant formula and increase the prevalence of breastfeeding from birth through 12 to 24 months of age through public health awareness campaigns. The specific goals and approaches of these breastfeeding promotion programs, and the policy environment surrounding their implementation, vary by country. As a policy basic framework, the International Code of Marketing of Breast-milk Substitutes, adopted by the WHO's World Health Assembly in 1981, requires infant formula companies to preface their product information with statements that breastfeeding is the best way of feeding babies and that a substitute should only be used after consultation with health professionals. The Baby Friendly Hospital Initiative also restricts use by hospitals of free formula or other infant care aids provided by formula companies. (See also Policy section below.)
Infant formula processing
History
Current general procedure
The manufacturing process may differ for different types of formula made; therefore the following is the general procedure for liquid-milk based formulas:
Mixing ingredients
Primary ingredients are blended in large stainless steel tanks and skim milk is added and adjusted to 60 °C. Then, fats, oils and emulsifiers are added. Additional heating and mixing may be required to get proper consistency. Next, minerals, vitamins, and stabilizing gums are added at various points, depending on their sensitivity to heat. The batch is temporarily stored and then transported by pipelines to pasteurization equipment when mixing is complete.
Pasteurization
This is a process that protects against spoilage by eliminating bacteria, yeasts and molds.
It involves quickly heating and then cooling of the product under controlled conditions which micro-organisms cannot survive. The batch is held at around 85–94 °C for approximately 30 seconds which is necessary to adequately reduce micro-organisms and prepare the formula for filling.
Homogenization
This is a process which increases emulsion uniformity and stability by reducing size of fat and oil particles in the formula. It is done with a variety of mixing equipment that applies shear to the product and this mixing breaks fat and oil particles into very small droplets.
Standardization
Standardization is used to ensure that the key parameters like pH, fat concentration and vitamins and mineral content are correct. If insufficient levels of these are found, the batch is reworked to achieve appropriate levels. After this step, the batch is ready to be packaged.
Packaging
Packaging depends on manufacturer and type of equipment used, but in general, liquid formula is filled into metal cans with lids crimped into place.
Heat treatment or sterilization
Finally, infant formulas are heat treated to maintain the bacteriologic quality of the product. This can be done traditionally by either retort sterilization or high-temperature short-time (HTST) treatment. Recently, ultrahigh-temperature treated formula has become more commonly used. If powdered formula is made, then spray drying would be required in addition. Retort sterilization is a traditional retort sterilization method that uses 10-15mins treatment at 118 °C. Ultrahigh-temperature (UHT) is a method that uses a brief (2–3 seconds) treatment at 142 °C. Because of the short time used, there is little protein denaturation, but the process still ensures sterility of the final product.
Recent and future potential new ingredients
Probiotics
Randomized, controlled trials completed in the 2000s have shown limited and short term clinical benefits for the use of probiotics in infants’ diet. The safety of probiotics in general and in infants, especially preterm infants, has been investigated in a limited number of controlled trials. The findings thus far suggest probiotics are generally safe, though the research is preliminary and has yet to provide definitive conclusions.
Prebiotics
Prebiotics are undigestible carbohydrates that promote the growth of probiotic bacteria in the gut. Human milk contains a variety of oligosaccharides believed to be an important factor in the pattern of microflora colonization of breastfed infants. Because of variety, variability, complexity and polymorphism of the oligosaccharide composition and structure, it is currently not feasible to reproduce the oligosaccharide components of human milk in a strictly structural fashion.
The European Society of Pediatric Gastroenterology, Hepatology, and Nutrition Committee on Nutrition found evidence to support short term effects of ingesting prebiotics on stool microflora of infants with increased in the number of bifidobacteria. Babies can be at risk of dehydration with the induction of softer stools, if they have the kidney immaturity and/or a poor ability to concentrate urine. A reduction of pathogens has been associated with the consumption of prebiotics. However, there was no evidence to support major clinical or long-term benefits. Therefore, there is little evidence of beneficial effects of prebiotics in dietary products.
Lysozyme and lactoferrin
Lysozyme is an enzyme that is responsible for protecting the body by damaging bacterial cell walls. Lactoferrin is a globular, multifunctional protein that has antimicrobial activity. Compared to human milk, cow's milk has a signifactly lower levels of lysozyme and lactoferrin; therefore, the industry has an increasing interest in adding them into infant formulas.
Long chain polyunsaturated fatty acid supplementation
Some manufacturers have begun supplementing formula milk with long-chain polyunsaturated fatty acids (LCPUFA). The current evidence suggests that there may be little or no difference between formula milk with and without LCPUFA supplementation in terms of babies' visual function, physical growth or neurodevelopment.
See also
2008 Chinese milk scandal
Baby food
Baby bottle
Breastfeeding
Breast milk
Child development
Daigou
Dairy allergy
List of dairy products
Baby Food Market Forecast to 2028 - Covid-19 Impact
Notes
References
External links
FDA 101: Infant Formula
"Breast-feeding and Guilt: Interview with a Mayo Clinic Specialist"
Infant and Toddler Nutrition
Breastfeeding VS Formula Feeding
Formula
Milk
Dairy products
Soy products
Bristol-Myers Squibb
Breast milk
Breastfeeding |
null | null | History of South Africa | eng_Latn | The first modern humans are believed to have inhabited South Africa more than 100,000 years ago. South Africa's prehistory has been divided into two phases based on broad patterns of technology namely the Stone Age and Iron Age. After the discovery of hominins at Taung and australopithecine fossils in limestone caves at Sterkfontein, Swartkrans, and Kromdraai these areas were collectively designated a World Heritage site. The first nations of South Africa are collectively referred to as the Khoisan, the Khoi Khoi and the San separately. These groups were displaced or sometimes absorbed by migrating Africans (Bantus) during the Bantu expansion from Western and Central Africa. While some maintained separateness, others were grouped into a category known as Coloureds, a multiracial ethnic group which includes people with shared ancestry from two or more of these groups: Khoisan, Bantu, English, Afrikaners, Austronesians, East Asians and South Asians. European exploration of the African coast began in the 13th century when Portugal committed itself to discover an alternative route to the silk road that would lead to China. In the 14th and 15th century, Portuguese explorers traveled down the west African Coast, detailing and mapping the coastline and in 1488 they rounded the Cape of Good Hope. The Dutch East India Company established a trading post in Cape Town under the command of Jan van Riebeeck in 1652, European workers who settled at the Cape became known as the Free Burghers and gradually established farms in the Dutch Cape Colony.
Following the Invasion of the Cape Colony in 1795 and 1806, mass migrations collectively known as the Great Trek occurred during which the Voortrekkers established several Boer settlements on the interior of South Africa. The discoveries of diamonds and gold in the nineteenth century had a profound effect on the fortunes of the region, propelling it onto the world stage and introducing a shift away from an exclusively agrarian-based economy towards industrialisation and the development of urban infrastructure. The discoveries also led to new conflicts culminating in open warfare between the Boer settlers and the British Empire, fought essentially for control over the nascent South African mining industry.
Following the defeat of the Boers in the Anglo–Boer or South African War (1899–1902), the Union of South Africa was created as a self-governing dominion of the British Empire on 31 May 1910 in terms of the South Africa Act 1909, which amalgamated the four previously separate British colonies: Cape Colony, Colony of Natal, Transvaal Colony, and Orange River Colony. The country became a fully sovereign nation state within the British Empire, in 1934 following enactment of the Status of the Union Act. The monarchy came to an end on 31 May 1961, replaced by a republic as the consequence of a 1960 referendum, which legitimised the country becoming the Republic of South Africa.
From 1948–1994, South African politics was dominated by Afrikaner nationalism. Racial segregation and white minority rule known officially as apartheid, an Afrikaans word meaning "separateness", was implemented in 1948. On 27 April 1994, after decades of armed struggle, terrorism and international opposition to apartheid, the African National Congress (ANC) achieved victory in the country's first democratic election. Since then, the African National Congress has governed South Africa, in an alliance with the South African Communist Party and the Congress of South African Trade Unions.
Early history (before 1652)
Prehistory
Scientists researching the periods before written historical records were made have established that the territory of what is now referred to generically as South Africa was one of the important centers of human evolution. It was inhabited by Australopithecines since at least 2.5 million years ago. Modern human settlement occurred around 125,000 years ago in the Middle Stone Age, as shown by archaeological discoveries at Klasies River Caves. The first human habitation is associated with a DNA group originating in a northwestern area of southern Africa and still prevalent in the indigenous Khoisan (Khoi and San). Southern Africa was later populated by Bantu-speaking people who migrated from the western region of central Africa during the early centuries AD.
At the Blombos cave Professor Raymond Dart discovered the skull of a 2.51 million year old Taung Child in 1924, the first example of Australopithecus africanus ever found. Following in Dart's footsteps Robert Broom discovered a new much more robust hominid in 1938 Paranthropus robustus at Kromdraai, and in 1947 uncovered several more examples of Australopithecus africanus at Sterkfontein. In further research at the Blombos cave in 2002, stones were discovered engraved with grid or cross-hatch patterns, dated to some 70,000 years ago. This has been interpreted as the earliest example ever discovered of abstract art or symbolic art created by Homo sapiens.
Many more species of early hominid have come to light in recent decades. The oldest is Little Foot, a collection of footbones of an unknown hominid between 2.2 and 3.3 million years old, discovered at Sterkfontein by Ronald J. Clarke. An important recent find was that of 1.9 million year old Australopithecus sediba, discovered in 2008. In 2015, the discovery near Johannesburg of a previously unknown species of Homo was announced, named Homo naledi. It has been described as one of the most important paleontological discoveries in modern times.
San and Khoikhoi
The descendants of the Middle Paleolithic populations are thought to be the aboriginal San and Khoikhoi tribes. These are collectively known as the Khoisan, a modern European portmanteau of these two tribes' names. The settlement of southern Africa by the Khoisan corresponds to the earliest separation of the extant Homo sapiens populations altogether, associated in genetic science with what is described in scientific terms as matrilinear haplogroup L0 (mtDNA) and patrilinear haplogroup A (Y-DNA), originating in a northwestern area of southern Africa.
The San and Khoikhoi are essentially distinguished only by their respective occupations. Whereas the San were hunter-gathers, the Khoikhoi were pastoral herders. The initial origin of the Khoikhoi remains uncertain.
Archaeological discoveries of livestock bones on the Cape Peninsula indicate that the Khoikhoi began to settle there by about 2000 years ago.
In the late 15th and early 16th centuries, Portuguese mariners, who were the first Europeans at the Cape, encountered pastoral Khoikhoi with livestock.
Later, English and Dutch seafarers in the late 16th and 17th centuries exchanged metals for cattle and sheep with the Khoikhoi.
The conventional view is that availability of livestock was one reason why, in the mid-17th century, the Dutch East India Company established a staging post where the port city of Cape Town is today situated.
The establishment of the staging post by the Dutch East India Company at the Cape in 1652 soon brought the Khoikhoi into conflict with Dutch settlers over land ownership. Cattle rustling and livestock theft ensued, with the Khoikhoi being ultimately expelled from the peninsula by force, after a succession of wars. The first Khoikhoi–Dutch War broke out in 1659, the second in 1673, and the third 1674–1677. By the time of their defeat and expulsion from the Cape Peninsula and surrounding districts, the Khoikhoi population was decimated by a smallpox epidemic introduced by Dutch sailors against which the Khoikhoi had no natural resistance or indigenous medicines.
The Bantu people
The Bantu expansion was one of the major demographic movements in human prehistory, sweeping much of the African continent during the 2nd and 1st millennia BC. Bantu-speaking communities reached southern Africa from the Congo basin as early as the 4th century BC. The advancing Bantu encroached on the Khoikhoi territory, forcing the original inhabitants of the region to move to more arid areas.
Some groups, ancestral to today's Nguni peoples (the Zulu, Xhosa, Swazi, and Ndebele), preferred to live near the eastern coast of what is present-day South Africa. Others, now known as the Sotho–Tswana peoples (Tswana, Pedi, and Sotho), settled in the interior on the plateau known as the Highveld, while today's Venda, Lemba, and Tsonga peoples made their homes in the north-eastern areas of present-day South Africa.
The Kingdom of Mapungubwe, which was located near the northern border of present-day South Africa, at the confluence of the Limpopo and Shashe rivers adjacent to present-day Zimbabwe and Botswana, was the first indigenous kingdom in southern Africa between AD 900 and 1300. It developed into the largest kingdom in the sub-continent before it was abandoned because of climatic changes in the 14th century. Smiths created objects of iron, copper and gold both for local decorative use and for foreign trade. The kingdom controlled trade through the east African ports to Arabia, India and China, and throughout southern Africa, making it wealthy through the exchange of gold and ivory for imports such as Chinese porcelain and Persian glass beads.
Specifics of the contact between Bantu-speakers and the indigenous Khoisan ethnic group remain largely unresearched, although linguistic proof of assimilation exists, as several southern Bantu languages (notably Xhosa and Zulu) are theorised in that they incorporate many click consonants from the Khoisan languages, as possibilities of such developing independently are valid as well.
Colonisation
Portuguese role
The Portuguese mariner Bartolomeu Dias was the first European to explore the coastline of South Africa in 1488, while attempting to discover a trade route to the Far East via the southernmost cape of South Africa, which he named Cabo das Tormentas, meaning Cape of Storms. In November 1497, a fleet of Portuguese ships under the command of the Portuguese mariner Vasco da Gama rounded the Cape of Good Hope. By 16 December, the fleet had passed the Great Fish River on the east coast of South Africa, where Dias had earlier turned back. Da Gama gave the name Natal to the coast he was passing, which in Portuguese means Christmas. Da Gama's fleet proceeded northwards to Zanzibar and later sailed eastwards, eventually reaching India and opening the Cape Route between Europe and Asia.
Dutch role
Dutch colonization (1652–1815)
The Dutch East India Company (in the Dutch of the day: Vereenigde Oostindische Compagnie, or VOC) decided to establish a permanent settlement at the Cape in 1652. The VOC, one of the major European trading houses sailing the spice route to the East, had no intention of colonizing the area, instead wanting only to establish a secure base camp where passing ships could be serviced and restock on supplies . To this end, a small VOC expedition under the command of Jan van Riebeeck reached Table Bay on 6 April 1652.
The VOC had settled at the Cape in order to supply their trading ships. The Khoikhoi stopped trading with the Dutch, and the Cape and the VOC had to import Dutch farmers to establish farms to supply the passing ships as well as to supply the growing VOC settlement. The small initial group of free burghers, as these farmers were known, steadily increased in number and began to expand their farms further north and east into the territory of the Khoikhoi. The free burghers were ex-VOC soldiers and gardeners, who were unable to return to Holland when their contracts were completed with the VOC. The VOC also brought some 71,000 slaves to Cape Town from India, Indonesia, East Africa, Mauritius, and Madagascar.
The majority of burghers had Dutch ancestry and belonged to the Dutch Reformed Church, but there were also some Germans, who often happened to be Lutherans. In 1688, the Dutch and the Germans were joined by French Huguenots, Calvinist Protestants fleeing religious persecution in France under its Catholic ruler, King Louis XIV.
Van Riebeeck considered it impolitic to enslave the local Khoi and San aboriginals, so the VOC began to import large numbers of slaves, primarily from the Dutch colonies in Indonesia. Eventually, van Riebeeck and the VOC began to make indentured servants out of the Khoikhoi and the San. The descendants of unions between the Dutch settlers and the Khoi-San and Malay slaves became known officially as the Cape Coloureds and the Cape Malays, respectively. A significant number of the offspring from the white and slave unions were absorbed into the local proto-Afrikaans speaking white population. The racially mixed genealogical origins of many so-called "white" South Africans have been traced to interracial unions at the Cape between the European occupying population and imported Asian and African slaves, the indigenous Khoi and San, and their vari-hued offspring. Simon van der Stel, the first Governor of the Dutch settlement, famous for his development of the lucrative South African wine industry, was himself of mixed race-origin.
British colonisation, Mfecane and Boer Republics (1815–1910)
British at the Cape
In 1787, shortly before the French Revolution, a faction within the politics of the Dutch Republic known as the Patriot Party attempted to overthrow the regime of stadtholder William V. Though the revolt was crushed, it was resurrected after the French invasion of the Netherlands in 1794/1795 which resulted in the stadtholder fleeing the country. The Patriot revolutionaries then proclaimed the Batavian Republic, which was closely allied to revolutionary France. In response, the stadtholder, who had taken up residence in England, issued the Kew Letters, ordering colonial governors to surrender to the British. The British then seized the Cape in 1795 to prevent it from falling into French hands. The Cape was relinquished back to the Dutch in 1803. In 1805, the British inherited the Cape as a prize during the Napoleonic Wars, again seizing the Cape from the French controlled Kingdom of Holland which had replaced the Batavian Republic.
Like the Dutch before them, the British initially had little interest in the Cape Colony, other than as a strategically located port. The Cape Articles of Capitulation of 1806 allowed the colony to retain "all their rights and privileges which they have enjoyed hitherto", and this launched South Africa on a divergent course from the rest of the British Empire, allowing the continuance of Roman-Dutch law. British sovereignty of the area was recognised at the Congress of Vienna in 1815, the Dutch accepting a payment of 6 million pounds for the colony. As one of their first tasks they outlawed the use of the Dutch language in 1806 with the view of converting the European settlers to the British language and culture. This had the effect of forcing more of the Dutch colonists to move (or trek) away from British administrative reach. Much later, in 1820 the British authorities persuaded about 5,000 middle-class British immigrants (most of them "in trade") to leave Great Britain. Many of the 1820 Settlers eventually settled in Grahamstown and Port Elizabeth.
British policy with regard to South Africa would vacillate with successive governments, but the overarching imperative throughout the 19th century was to protect the strategic trade route to India while incurring as little expense as possible within the colony. This aim was complicated by border conflicts with the Boers, who soon developed a distaste for British authority.
European exploration of the interior
Colonel Robert Jacob Gordon of the Dutch East India Company was the first European to explore parts of the interior while commanding the Dutch garrison at the renamed Cape of Good Hope, from 1780 to 1795. The four expeditions Gordon undertook between 1777 and 1786 are recorded in a series of several hundred drawings known collectively as the Gordon Atlas, as well as in his journals, which were only discovered in 1964.
Early relations between the European settlers and the Xhosa, the first Bantu peoples they met when they ventured inward, were peaceful. However, there was competition for land, and this tension led to skirmishes in the form of cattle raids from 1779.
The British explorers David Livingstone and William Oswell, setting out from a mission station in the northern Cape Colony, are believed to have been the first white men to cross the Kalahari desert in 1849. The Royal Geographical Society later awarded Livingstone a gold medal for his discovery of Lake Ngami in the desert.
Zulu militarism and expansionism
The Zulu people are part of the Nguni tribe and were originally a minor clan in what is today northern KwaZulu-Natal, founded ca. 1709 by Zulu kaNtombela.
The 1820s saw a time of immense upheaval relating to the military expansion of the Zulu Kingdom, which replaced the original African clan system with kingdoms. Sotho-speakers know this period as the difaqane ("forced migration"); Zulu-speakers call it the mfecane ("crushing").
Various theories have been advanced for the causes of the difaqane, ranging from ecological factors to competition in the ivory trade. Another theory attributes the epicentre of Zulu violence to the slave trade out of Delgoa Bay in Mozambique situated to the north of Zululand. Most historians recognise that the Mfecane wasn't just a series of events caused by the founding of the Zulu kingdom but rather a multitude of factors caused before and after Shaka Zulu came into power.
In 1818, Nguni tribes in Zululand created a militaristic kingdom between the Tugela River and Pongola River, under the driving force of Shaka kaSenzangakhona, son of the chief of the Zulu clan. Shaka built large armies, breaking from clan tradition by placing the armies under the control of his own officers rather than of hereditary chiefs. He then set out on a massive programme of expansion, killing or enslaving those who resisted in the territories he conquered. His impis (warrior regiments) were rigorously disciplined: failure in battle meant death.
The Zulu resulted in the mass movement of many tribes who in turn to tried dominate those in new territories, leading to widespread warfare and waves of displacement spread throughout southern Africa and beyond. It accelerated the formation of several new nation-states, notably those of the Sotho (present-day Lesotho) and the Swazi (now Eswatini (formerly Swaziland)). It caused the consolidation of groups such as the Matebele, the Mfengu and the Makololo.
In 1828 Shaka was killed by his half-brothers Dingaan and Umhlangana. The weaker and less-skilled Dingaan became king, relaxing military discipline while continuing the despotism. Dingaan also attempted to establish relations with the British traders on the Natal coast, but events had started to unfold that would see the demise of Zulu independence. Estimates for the death toll resulting from the Mfecane range from 1 million to 2 million.
Boer people and republics
After 1806, a number of Dutch-speaking inhabitants of the Cape Colony trekked inland, first in small groups. Eventually, in the 1830s, large numbers of Boers migrated in what came to be known as the Great Trek. Among the initial reasons for their leaving the Cape colony were the English language rule. Religion was a very important aspect of the settlers culture and the bible and church services were in Dutch. Similarly, schools, justice and trade up to the arrival of the British, were all managed in the Dutch language. The language law caused friction, distrust and dissatisfaction.
Another reason for Dutch-speaking white farmers trekking away from the Cape was the abolition of slavery by the British government on Emancipation Day, 1 December 1838. The farmers complained they could not replace the labour of their slaves without losing an excessive amount of money. The farmers had invested large amounts of capital in slaves. Owners who had purchased slaves on credit or put them up as surety against loans faced financial ruin. Britain had allocated the sum of 1 200 000 British Pounds as compensation to the Dutch settlers, on condition the Dutch farmers had to lodge their claims in Britain as well as the fact that the value of the slaves was many times the allocated amount. This caused further dissatisfaction among the Dutch settlers. The settlers, incorrectly, believed that the Cape Colony administration had taken the money due to them as payment for freeing their slaves. Those settlers who were allocated money could only claim it in Britain in person or through an agent. The commission charged by agents was the same as the payment for one slave, thus those settlers only claiming for one slave would receive nothing.
South African Republic
The South African Republic (Dutch: Zuid-Afrikaansche Republiek or ZAR, not to be confused with the much later Republic of South Africa), is often referred to as The Transvaal and sometimes as the Republic of Transvaal. It was an independent and internationally recognised nation-state in southern Africa from 1852 to 1902. Independent sovereignty of the republic was formally recognised by Great Britain with the signing of the Sand River Convention on 17 January 1852. The republic, under the premiership of Paul Kruger, defeated British forces in the First Boer War and remained independent until the end of the Second Boer War on 31 May 1902, when it was forced to surrender to the British. The territory of the South African Republic became known after this war as the Transvaal Colony.
Free State Republic
The independent Boer republic of Orange Free State evolved from colonial Britain's Orange River Sovereignty, enforced by the presence of British troops, which lasted from 1848 to 1854 in the territory between the Orange and Vaal rivers, named Transorange. Britain, due to the military burden imposed on it by the Crimean War in Europe, then withdrew its troops from the territory in 1854, when the territory along with other areas in the region was claimed by the Boers as an independent Boer republic, which they named the Orange Free State. In March 1858, after land disputes, cattle rustling and a series of raids and counter-raids, the Orange Free State declared war on the Basotho kingdom, which it failed to defeat. A succession of wars were conducted between the Boers and the Basotho for the next 10 years. The name Orange Free State was again changed to the Orange River Colony, created by Britain after the latter occupied it in 1900 and then annexed it in 1902 during the Second Boer War. The colony, with an estimated population of less than 400,000 in 1904 ceased to exist in 1910, when it was absorbed into the Union of South Africa as the Orange Free State Province.
Natalia
Natalia was a short-lived Boer republic established in 1839 by Boer Voortrekkers emigrating from the Cape Colony. In 1824 a party of 25 men under British Lieutenant F G Farewell arrived from the Cape Colony and established a settlement on the northern shore of the Bay of Natal, which would later become the port of Durban, so named after Benjamin D'Urban, a governor of the Cape Colony. Boer Voortrekkers in 1838 established the Republic of Natalia in the surrounding region, with its capital at Pietermaritzburg. On the night of 23/24 May 1842 British colonial forces attacked the Voortrekker camp at Congella. The attack failed, with British forces then retreating back to Durban, which the Boers besieged. A local trader Dick King and his servant Ndongeni, who later became folk heroes, were able to escape the blockade and ride to Grahamstown, a distance of 600 km (372.82 miles) in 14 days to raise British reinforcements. The reinforcements arrived in Durban 20 days later; the siege was broken and the Voortrekkers retreated. The Boers accepted British annexation in 1844. Many of the Natalia Boers who refused to acknowledge British rule trekked over the Drakensberg mountains to settle in the Orange Free State and Transvaal republics.
Cape Colony
Between 1847 and 1854, Harry Smith, governor and high commissioner of the Cape Colony, annexed territories far to the north of original British and Dutch settlement.
Smith's expansion of the Cape Colony resulted in conflict with disaffected Boers in the Orange River Sovereignty who in 1848 mounted an abortive rebellion at Boomplaats, where the Boers were defeated by a detachment of the Cape Mounted Rifles. Annexation also precipitated a war between British colonial forces and the indigenous Xhosa nation in 1850, in the eastern coastal region.
Starting from the mid-1800s, the Cape of Good Hope, which was then the largest state in southern Africa, began moving towards greater independence from Britain.
In 1854, it was granted its first locally elected legislature, the Cape Parliament.
In 1872, after a long political struggle, it attained responsible government with a locally accountable executive and Prime Minister. The Cape nonetheless remained nominally part of the British Empire, even though it was self-governing in practice.
The Cape Colony was unusual in southern Africa in that its laws prohibited any discrimination on the basis of race and, unlike the Boer republics, elections were held according to the non-racial Cape Qualified Franchise system, whereby suffrage qualifications applied universally, regardless of race.
Initially, a period of strong economic growth and social development ensued. However, an ill-informed British attempt to force the states of southern Africa into a British federation led to inter-ethnic tensions and the First Boer War. Meanwhile, the discovery of diamonds around Kimberley and gold in the Transvaal led to a later return to instability, particularly because they fueled the rise to power of the ambitious colonialist Cecil Rhodes. As Cape Prime Minister, Rhodes curtailed the multi-racial franchise, and his expansionist policies set the stage for the Second Boer War.
Natal
Indian slaves from the Dutch colonies in India had been introduced into the Cape area of South Africa by the Dutch settlers in 1654.
By the end of 1847, following annexation by Britain of the former Boer republic of Natalia, nearly all the Boers had left their former republic, which the British renamed Natal. The role of the Boer settlers was replaced by subsidised British immigrants of whom 5,000 arrived between 1849 and 1851.
By 1860, with slavery having been abolished in 1834, and after the annexation of Natal as a British colony in 1843, the British colonists in Natal (now kwaZulu-Natal) turned to India to resolve a labour shortage, as men of the local Zulu warrior nation were refusing to work on the plantations and farms established by the colonists. In that year, the SS Truro arrived in Durban harbour with over 300 Indians on board.
Over the next 50 years, 150,000 more indentured Indian servants and labourers arrived, as well as numerous free "passenger Indians," building the base for what would become the largest Indian diasporic community outside India.
By 1893, when the lawyer and social activist Mahatma Gandhi arrived in Durban, Indians outnumbered whites in Natal. The civil rights struggle of Gandhi's Natal Indian Congress failed; until the 1994 advent of democracy, Indians in South Africa were subject to most of the discriminatory laws that applied to all non-white inhabitants of the country.
Griqua people
By the late 1700s, the Cape Colony population had grown to include a large number of mixed-race so-called "coloureds" who were the offspring of extensive interracial relations between male Dutch settlers, Khoikhoi females, and female slaves imported from Dutch colonies in the East. Members of this mixed-race community formed the core of what was to become the Griqua people.
Under the leadership of a former slave named Adam Kok, these "coloureds" or Basters (meaning mixed race or multiracial) as they were named by the Dutch—a word derived from baster, meaning "bastard"—started trekking northward into the interior, through what is today named Northern Cape Province. The trek of the Griquas to escape the influence of the Cape Colony has been described as "one of the great epics of the 19th century." They were joined on their long journey by a number of San and Khoikhoi aboriginals, local African tribesmen, and also some white renegades. Around 1800, they started crossing the northern frontier formed by the Orange River, arriving ultimately in an uninhabited area, which they named Griqualand.
In 1825, a faction of the Griqua people was induced by Dr John Philip, superintendent of the London Missionary Society in Southern Africa, to relocate to a place called Philippolis, a mission station for the San, several hundred miles southeast of Griqualand. Philip's intention was for the Griquas to protect the missionary station there against banditti in the region, and as a bulwark against the northward movement of white settlers from the Cape Colony. Friction between the Griquas and the settlers over land rights resulted in British troops being sent to the region in 1845. It marked the beginning of nine years of British intervention in the affairs of the region, which the British named Transorange.
In 1861, to avoid the imminent prospect of either being colonised by the Cape Colony or coming into conflict with the expanding Boer Republic of Orange Free State, most of the Philippolis Griquas embarked on a further trek. They moved about 500 miles eastward, over the Quathlamba (today known as the Drakensberg mountain range), settling ultimately in an area officially designated as "Nomansland", which the Griquas renamed Griqualand East. East Griqualand was subsequently annexed by Britain in 1874 and incorporated into the Cape Colony in 1879.
The original Griqualand, north of the Orange River, was annexed by Britain's Cape Colony and renamed Griqualand West after the discovery in 1871 of the world's richest deposit of diamonds at Kimberley, so named after the British Colonial Secretary, Earl Kimberley.
Although no formally surveyed boundaries existed, Griqua leader Nicolaas Waterboer claimed the diamond fields were situated on land belonging to the Griquas. The Boer republics of Transvaal and the Orange Free State also vied for ownership of the land, but Britain, being the preeminent force in the region, won control over the disputed territory. In 1878, Waterboer led an unsuccessful rebellion against the colonial authorities, for which he was arrested and briefly exiled.
Wars against the Xhosa
In early South Africa, European notions of national boundaries and land ownership had no counterparts in African political culture. To Moshoeshoe the BaSotho chieftain from Lesotho, it was customary tribute in the form of horses and cattle represented acceptance of land use under his authority. To European settlers in Southern Africa, the same form of tribute was believed to constitute purchase and permanent ownership of the land under independent authority.
As European settlers started establishing permanent farms after trekking across the country in search of prime agricultural land, they encountered resistance from the local Bantu people who had originally migrated southwards from central Africa hundreds of years earlier. The consequent frontier wars became known as the Xhosa Wars (which were also referred to in contemporary discussion as the Kafir Wars or the Cape Frontier Wars). In the southeastern part of South Africa, Boer settlers and the Xhosa clashed along the Great Fish River, and in 1779 the First Xhosa War broke out. For nearly 100 years subsequently, the Xhosa fought the settlers sporadically, first the Boers or Afrikaners and later the British. In the Fourth Xhosa War, which lasted from 1811 to 1812, the British colonial authorities forced the Xhosa back across the Great Fish River and established forts along this boundary.
The increasing economic involvement of the British in southern Africa from the 1820s, and especially following the discovery of first diamonds at Kimberley and gold in the Transvaal, resulted in pressure for land and African labour, and led to increasingly tense relations with Southern African states.
In 1818 differences between two Xhosa leaders, Ndlambe and Ngqika, ended in Ngqika's defeat, but the British continued to recognise Ngqika as the paramount chief. He appealed to the British for help against Ndlambe, who retaliated in 1819 during the Fifth Frontier War by attacking the British colonial town of Grahamstown.
Wars against the Zulu
In the eastern part of what is today South Africa, in the region named Natalia by the Boer trekkers, the latter negotiated an agreement with Zulu King Dingane kaSenzangakhona allowing the Boers to settle in part of the then Zulu kingdom. Cattle rustling ensued and a party of Boers under the leadership of Piet Retief were killed.
Subsequent to the killing of the Retief party, the Boers defended themselves against a Zulu attack, at the Ncome River on 16 December 1838. An estimated five thousand Zulu warriors were involved. The Boers took a defensive position with the high banks of the Ncome River forming a natural barrier to their rear with their ox waggons as barricades between themselves and the attacking Zulu army. About three thousand Zulu warriors died in the clash known historically as the Battle of Blood River.
In the later annexation of the Zulu kingdom by imperial Britain, an Anglo-Zulu War was fought in 1879. Following Lord Carnarvon's successful introduction of federation in Canada, it was thought that similar political effort, coupled with military campaigns, might succeed with the African kingdoms, tribal areas and Boer republics in South Africa.
In 1874, Henry Bartle Frere was sent to South Africa as High Commissioner for the British Empire to bring such plans into being. Among the obstacles were the presence of the independent states of the South African Republic and the Kingdom of Zululand and its army. Frere, on his own initiative, without the approval of the British government and with the intent of instigating a war with the Zulu, had presented an ultimatum on 11 December 1878, to the Zulu king Cetshwayo with which the Zulu king could not comply. Bartle Frere then sent Lord Chelmsford to invade Zululand. The war is notable for several particularly bloody battles, including an overwhelming victory by the Zulu at the Battle of Isandlwana, as well as for being a landmark in the timeline of imperialism in the region.
Britain's eventual defeat of the Zulus, marking the end of the Zulu nation's independence, was accomplished with the assistance of Zulu collaborators who harboured cultural and political resentments against centralised Zulu authority. The British then set about establishing large sugar plantations in the area today named KwaZulu-Natal Province.
Wars with the Basotho
From the 1830s onwards, numbers of white settlers from the Cape Colony crossed the Orange River and started arriving in the fertile southern part of territory known as the Lower Caledon Valley, which was occupied by Basotho cattle herders under the authority of the Basotho founding monarch Moshoeshoe I. In 1845, a treaty was signed between the British colonists and Moshoeshoe, which recognised white settlement in the area. No firm boundaries were drawn between the area of white settlement and Moshoeshoe's kingdom, which led to border clashes. Moshoeshoe was under the impression he was loaning grazing land to the settlers in accordance with African precepts of occupation rather than ownership, while the settlers believed they had been granted permanent land rights. Afrikaner settlers in particular were loath to live under Moshoesoe's authority and among Africans.
The British, who at that time controlled the area between the Orange and Vaal Rivers called the Orange River Sovereignty, decided a discernible boundary was necessary and proclaimed a line named the Warden Line, dividing the area between British and Basotho territories. This led to conflict between the Basotho and the British, who were defeated by Moshoeshoe's warriors at the battle of Viervoet in 1851.
As punishment to the Basotho, the governor and commander-in-chief of the Cape Colony, George Cathcart, deployed troops to the Mohokare River; Moshoeshoe was ordered to pay a fine. When he did not pay the fine in full, a battle broke out on the Berea Plateau in 1852, where the British suffered heavy losses. In 1854, the British handed over the territory to the Boers through the signing of the Sand River Convention. This territory and others in the region then became the Republic of the Orange Free State.
A succession of wars followed from 1858 to 1868 between the Basotho kingdom and the Boer republic of Orange Free State. In the battles that followed, the Orange Free State tried unsuccessfully to capture Moshoeshoe's mountain stronghold at Thaba Bosiu, while the Sotho conducted raids in Free State territories. Both sides adopted scorched-earth tactics, with large swathes of pasturage and cropland being destroyed. Faced with starvation, Moshoeshoe signed a peace treaty on 15 October 1858, though crucial boundary issues remained unresolved. War broke out again in 1865. After an unsuccessful appeal for aid from the British Empire, Moshoeshoe signed the 1866 treaty of Thaba Bosiu, with the Basotho ceding substantial territory to the Orange Free State. On 12 March 1868, the British parliament declared the Basotho Kingdom a British protectorate and part of the British Empire. Open hostilities ceased between the Orange Free State and the Basotho. The country was subsequently named Basutoland and is presently named Lesotho.
Wars with the Ndebele
In 1836, when Boer Voortrekkers (pioneers) arrived in the northwestern part of present-day South Africa, they came into conflict with a Ndebele sub-group that the settlers named "Matabele", under chief Mzilikazi. A series of battles ensued, in which Mzilikazi was eventually defeated. He withdrew from the area and led his people northwards to what would later become the Matabele region of Southern Rhodesia (now Zimbabwe).
Other members of the Ndebele ethnic language group in different areas of the region similarly came into conflict with the Voortrekkers, notably in the area that would later become the Northern Transvaal. In September 1854, 28 Boers accused of cattle rustling were killed in three separate incidents by an alliance of the Ndebele chiefdoms of Mokopane and Mankopane. Mokopane and his followers, anticipating retaliation by the settlers, retreated into the mountain caves known as Gwasa, (or Makapansgat in Afrikaans). In late October, Boer commandos supported by local Kgatla tribal collaborators laid siege to the caves. By the end of the siege, about three weeks later, Mokopane and between 1,000 and 3,000 people had died in the caves. The survivors were captured and allegedly enslaved.
Wars with the Bapedi
The Bapedi wars, also known as the Sekhukhune wars, consisted of three separate campaigns fought between 1876 and 1879 against the Bapedi under their reigning monarch King Sekhukhune I, in the northeastern region known as Sekhukhuneland, bordering on Swaziland. Further friction was caused by the refusal of Sekhukhune to allow prospectors to search for gold in territory he considered to be sovereign and independent under his authority. The First Sekhukhune War of 1876 was conducted by the Boers, and the two separate campaigns of the Second Sekhukhune War of 1878/1879 were conducted by the British.
During the final campaign, Sekukuni (also spelled Sekhukhune) and members of his entourage took refuge in a mountain cave where he was cut off from food and water. He eventually surrendered to a combined deputation of Boer and British forces on 2 December 1879. Sekhukhune, members of his family and some Bapedi generals were subsequently imprisoned in Pretoria for two years, with Sekhukhuneland becoming part of the Transvaal Republic. No gold was ever discovered in the annexed territory.
Discovery of diamonds
The first diamond discoveries between 1866 and 1867 were alluvial, on the southern banks of the Orange River. By 1869, diamonds were found at some distance from any stream or river, in hard rock called blue ground, later called kimberlite, after the mining town of Kimberley where the diamond diggings were concentrated. The diggings were located in an area of vague boundaries and disputed land ownership. Claimants to the site included the South African (Transvaal) Republic, the Orange Free State Republic, and the mixed-race Griqua nation under Nicolaas Waterboer. Cape Colony Governor Henry Barkly persuaded all claimants to submit themselves to a decision of an arbitrator and so Robert W Keate, Lieutenant-Governor of Natal was asked to arbitrate. Keate awarded ownership to the Griquas. Waterboer, fearing conflict with the Boer republic of Orange Free State, subsequently asked for and received British protection. Griqualand then became a separate Crown Colony renamed Griqualand West in 1871, with a Lieutenant-General and legislative council.
The Crown Colony of Griqualand West was annexed into the Cape Colony in 1877, enacted into law in 1880. No material benefits accrued to the Griquas as a result of either colonisation or annexation; they did not receive any share of the diamond wealth generated at Kimberley. The Griqua community became subsequently dissimulated.
By the 1870s and 1880s the mines at Kimberley were producing 95% of the world's diamonds. The widening search for gold and other resources were financed by the wealth produced and the practical experience gained at Kimberley. Revenue accruing to the Cape Colony from the Kimberley diamond diggings enabled the Cape Colony to be granted responsible government status in 1872, since it was no longer dependent on the British Treasury and hence allowing it to be fully self-governing in similar fashion to the federation of Canada, New Zealand and some of the Australian states. The wealth derived from Kimberley diamond mining, having effectively tripled the customs revenue of the Cape Colony from 1871 to 1875, also doubled its population, and allowed it to expand its boundaries and railways to the north.
In 1888, British mining magnate Cecil John Rhodes co-founded De Beers Consolidated Mines at Kimberley, after buying up and amalgamating the individual claims with finance provided by the Rothschild dynasty. Abundant, cheap African labour was central to the success of Kimberley diamond mining, as it would later also be to the success of gold mining on the Witwatersrand. It has been suggested in some academic circles that the wealth produced at Kimberley was a significant factor influencing the Scramble for Africa, in which European powers had by 1902 competed with each other in drawing arbitrary boundaries across almost the entire continent and dividing it among themselves.
Discovery of gold
Although many tales abound, there is no conclusive evidence as to who first discovered gold or the manner in which it was originally discovered in the late 19th century on the Witwatersrand (meaning White Waters Ridge) of the Transvaal. The discovery of gold in February 1886 at a farm called Langlaagte on the Witwatersrand in particular precipitated a gold rush by prospectors and fortune seekers from all over the world. Except in rare outcrops, however, the main gold deposits had over many years become covered gradually by thousands of feet of hard rock. Finding and extracting the deposits far below the ground called for the capital and engineering skills that would soon result in the deep-level mines of the Witwatersrand producing a quarter of the world's gold, with the "instant city" of Johannesburg arising astride the main Witwatersrand gold reef.
Within two years of gold being discovered on the Witwatersrand, four mining finance houses had been established. The first was formed by Hermann Eckstein in 1887, eventually becoming Rand Mines. Cecil Rhodes and Charles Rudd followed, with their Gold Fields of South Africa company. Rhodes and Rudd had earlier made fortunes from diamond mining at Kimberley. In 1895 there was an investment boom in Witwatersrand gold-mining shares. The precious metal that underpinned international trade would dominate South African exports for decades to come.
Of the leading 25 foreign industrialists who were instrumental in opening up deep level mining operations at the Witwatersrand gold fields, 15 were Jewish, 11 of the total were from Germany or Austria, and nine of that latter category were also Jewish. The commercial opportunities opened by the discovery of gold attracted many other people of European Jewish origin. The Jewish population of South Africa in 1880 numbered approximately 4,000; by 1914 it had grown to more than 40,000, mostly migrants from Lithuania.
The working environment of the mines, meanwhile, as one historian has described it, was "dangerous, brutal and onerous", and therefore unpopular among local black Africans. Recruitment of black labour began to prove difficult, even with an offer of improved wages. In mid-1903 there remained barely half of the 90,000 black labourers who had been employed in the industry in mid-1899. The decision was made to start importing Chinese indentured labourers who were prepared to work for far less wages than local African labourers. The first 1,000 indentured Chinese labourers arrived in June 1904. By January 1907, 53,000 Chinese labourers were working in the gold mines.
First Anglo–Boer War
The Transvaal Boer republic was forcefully annexed by Britain in 1877, during Britain's attempt to consolidate the states of southern Africa under British rule. Long-standing Boer resentment turned into full-blown rebellion in the Transvaal and the first Anglo–Boer War, also known as the Boer Insurrection, broke out in 1880. The conflict ended almost as soon as it began with a decisive Boer victory at Battle of Majuba Hill (27 February 1881).
The republic regained its independence as the Zuid-Afrikaansche Republiek ("South African Republic"), or ZAR. Paul Kruger, one of the leaders of the uprising, became President of the ZAR in 1883. Meanwhile, the British, who viewed their defeat at Majuba as an aberration, forged ahead with their desire to federate the Southern African colonies and republics. They saw this as the best way to come to terms with the fact of a white Afrikaner majority, as well as to promote their larger strategic interests in the area.
The cause of the Anglo–Boer wars has been attributed to a contest over which nation would control and benefit most from the Witwatersrand gold mines. The enormous wealth of the mines was in the hands of European "Randlords" overseeing the mainly British foreign managers, mining foremen, engineers and technical specialists, characterised by the Boers as uitlander, meaning aliens. The "aliens" objected to being denied parliamentary representation and the right to vote, and they complained also of bureaucratic government delays in the issuing of licenses and permits, and general administrative incompetence on the part of the government.
In 1895, a column of mercenaries in the employ of Cecil John Rhodes' Rhodesian-based Charter Company and led by Captain Leander Starr Jameson had entered the ZAR with the intention of sparking an uprising on the Witwatersrand and installing a British administration there. The armed incursion became known as the Jameson Raid. It ended when the invading column was ambushed and captured by Boer commandos. President Kruger suspected the insurgency had received at least the tacit approval of the Cape Colony government under the premiership of Cecil John Rhodes, and that Kruger's South African Republic faced imminent danger. Kruger reacted by forming an alliance with the neighbouring Boer republic of Orange Free State. This did not prevent the outbreak of a Second Anglo–Boer war.
Second Anglo–Boer War
Renewed tensions between Britain and the Boers peaked in 1899 when the British demanded voting rights for the 60,000 foreign whites on the Witwatersrand. Until that point, President Paul Kruger's government had excluded all foreigners from the franchise. Kruger rejected the British demand and called for the withdrawal of British troops from the borders of the South African Republic. When the British refused, Kruger declared war. This Second Anglo–Boer War, also known as the South African War lasted longer than the first, with British troops being supplemented by colonial troops from Southern Rhodesia, Canada, India, Australia and New Zealand. It has been estimated that the total number of British and colonial troops deployed in South Africa during the war outnumbered the population of the two Boer Republics by more than 150,000.
By June 1900, Pretoria, the last of the major Boer towns, had surrendered. Yet resistance by Boer bittereinders (meaning those who would fight to the bitter end) continued for two more years with guerrilla warfare, which the British met in turn with scorched earth tactics. The Boers kept on fighting.
The British suffragette Emily Hobhouse visited British concentration camps in South Africa and produced a report condemning the appalling conditions there. By 1902, 26,000 Boer women and children had died of disease and neglect in the camps.
The Anglo–Boer War affected all race groups in South Africa. Black people were conscripted or otherwise coerced by both sides into working for them either as combatants or non-combatants to sustain the respective war efforts of both the Boers and the British. The official statistics of blacks killed in action are inaccurate. Most of the bodies were dumped in unmarked graves. It has, however, been verified that 17,182 black people died mainly of diseases in the Cape concentration camps alone, but this figure is not accepted historically as a true reflection of the overall numbers. Concentration camp superintendents did not always record the deaths of black inmates in the camps.
From the outset of hostilities in October 1899 to the signing of peace on 31 May 1902 the war claimed the lives of 22,000 imperial soldiers and 7,000 republican fighters. In terms of the peace agreement known as the Treaty of Vereeniging, the Boer republics acknowledged British sovereignty, while the British in turn committed themselves to reconstruction of the areas under their control.
Union of South Africa (1910–1948)
During the years immediately following the Anglo–Boer wars, Britain set about unifying the four colonies including the former Boer republics into a single self-governed country called the Union of South Africa. This was accomplished after several years of negotiations, when the South Africa Act 1909 consolidated the Cape Colony, Natal, Transvaal, and Orange Free State into one nation. Under the provisions of the act, the Union became an independent Dominion of the British Empire, governed under a form of constitutional monarchy, with the British monarch represented by a Governor-General. Prosecutions before the courts of the Union of South Africa were instituted in the name of the Crown and government officials served in the name of the Crown. The British High Commission territories of Basutoland (now Lesotho), Bechuanaland (now Botswana), and Swaziland (now Eswatini) continued under direct rule from Britain.
Among other harsh segregationist laws, including denial of voting rights to black people, the Union parliament enacted the 1913 Natives' Land Act, which earmarked only eight percent of South Africa's available land for black occupancy. White people, who constituted 20 percent of the population, held 90 percent of the land. The Land Act would form a cornerstone of legalised racial discrimination for the next nine decades.
General Louis Botha headed the first government of the new Union, with General Jan Smuts as his deputy. Their South African National Party, later known as the South African Party or SAP, followed a generally pro-British, white-unity line. The more radical Boers split away under the leadership of General Barry Hertzog, forming the National Party (NP) in 1914. The National Party championed Afrikaner interests, advocating separate development for the two white groups, and independence from Britain.
Dissatisfaction with British influence in the Union's affairs reached a climax in September 1914, when impoverished Boers, anti-British Boers and bitter-enders launched a rebellion. The rebellion was suppressed, and at least one officer was sentenced to death and executed by firing squad.
In 1924 the Afrikaner-dominated National Party came to power in a coalition government with the Labour Party. Afrikaans, previously regarded as a low-level Dutch patois, replaced Dutch as an official language of the Union. English and Dutch became the two official languages in 1925.
The Union of South Africa came to an end after a referendum on 5 October 1960, in which a majority of white South Africans voted in favour of unilateral withdrawal from the British Commonwealth and the establishment of a Republic of South Africa.
First World War
At the outbreak of World War I, South Africa joined Great Britain and the Allies against the German Empire. Both Prime Minister Louis Botha and Defence Minister Jan Smuts were former Second Boer War generals who had previously fought against the British, but they now became active and respected members of the Imperial War Cabinet. Elements of the South African Army refused to fight against the Germans and along with other opponents of the government; they rose in an open revolt known as the Maritz Rebellion. The government declared martial law on 14 October 1914, and forces loyal to the government under the command of generals Louis Botha and Jan Smuts defeated the rebellion. The rebel leaders were prosecuted, fined heavily and sentenced to imprisonment ranging from six to seven years.
Public opinion in South Africa split along racial and ethnic lines. The British elements strongly supported the war, and formed by far the largest military component. Likewise the Indian element (led by Mahatma Gandhi) generally supported the war effort. Afrikaners were split, with some like Botha and Smuts taking a prominent leadership role in the British war effort. This position was rejected by many rural Afrikaners who supported the Maritz Rebellion. The trade union movement was divided. Many urban blacks supported the war expecting it would raise their status in society. Others said it was not relevant to the struggle for their rights. The Coloured element was generally supportive and many served in a Coloured Corps in East Africa and France, also hoping to better themselves after the war.
With a population of roughly 6 million, between 1914–1918, over 250,000 South Africans of all races voluntarily served their country. Thousands more served in the British Army directly, with over 3,000 joining the British Royal Flying Corps and over 100 volunteering for the Royal Navy. It is likely that around 50% of white men of military age served during the war, more than 146,000 whites. 83,000 Blacks and 2,500 Coloureds and Asians also served in either German South-West Africa, East Africa, the Middle East, or on the Western Front in Europe. Over 7,000 South Africans were killed, and nearly 12,000 were wounded during the course of the war. Eight South Africans won the Victoria Cross for gallantry, the Empire's highest and prestigious military medal. The Battle of Delville Wood and the sinking of the SS Mendi being the greatest single incidents of loss of life.
25,000 Black South Africans were recruited at the request of the British War Cabinet to serve as non-combatant labourers in the South African Native Labour Contingent (SANLC). 21,000 of them were deployed to France as stevedores at French ports, where they were housed in segregated compounds. A total of 616 men from the Fifth Battalion of the SANLC drowned on 21 February 1917 when the troopship SS Mendi, on which they were being transported to France, collided with another vessel near the Isle of Wight. The Mendi disaster was one of South Africa's worst tragedies of the Great War, second perhaps only to the Battle of Delville Wood. The South African government issued no war service medal to the black servicemen and the special medal issued by King George V to "native troops" that served the Empire, the British War Medal in bronze, was disallowed and not issued to the SANLC.
Black and mixed-race South Africans who had supported the war were embittered when post-war South Africa saw no easing of white domination and racial segregation.
The assistance that South Africa gave the British Empire was significant. Two German African colonies were occupied, either by South Africa alone or with significant South African assistance. Manpower, from all races, helped Allied operations not just on the Western Front and Africa, but also in the Middle East against the Ottoman Empire. South Africa's ports and harbours on the Home Front were a crucial strategic asset when conducting a war on a global scale. Providing important rest and refuelling stations, the Royal Navy could ensure vital sea lane connections to the British Raj, and the Far East stayed open.
Economically, South Africa supplied two-thirds of gold production in the British Empire, with most of the remainder coming from Australia. At the start of the war, Bank of England officials in London worked with South Africa to block gold shipments to Germany, and force mine owners to sell only to the British Treasury, at prices set by the Treasury. This facilitated purchases of munitions and food in the United States and neutral countries.
Second World War
During World War II, South Africa's ports and harbours, such as at Cape Town, Durban, and Simon's Town, were important strategic assets to the British Royal Navy. South Africa's top-secret Special Signals Service played a significant role in the early development and deployment of radio detection and ranging (radar) technology used in protecting the vital coastal shipping route around southern Africa. By August 1945, South African Air Force aircraft in conjunction with British and Dutch aircraft stationed in South Africa had intercepted 17 enemy ships, assisted in the rescue of 437 survivors of sunken ships, attacked 26 of the 36 enemy submarines operating the vicinity of the South African coast, and flown 15,000 coastal patrol sorties.
About 334,000 South Africans volunteered for full-time military service in support of the Allies abroad. Nearly 9,000 were killed in action. On 21 June 1942 nearly 10,000 South African soldiers, representing one-third of the entire South African force in the field, were taken prisoner by German Field Marshal Rommel's forces in the fall of Tobruk, Libya. A number of South African fighter pilots served with distinction in the Royal Air Force during the Battle of Britain, including Group Captain Adolph "Sailor" Malan who led 74 Squadron and established a record of personally destroying 27 enemy aircraft.
General Jan Smuts was the only important non-British general whose advice was constantly sought by Britain's war-time Prime Minister Winston Churchill. Smuts was invited to the Imperial War Cabinet in 1939 as the most senior South African in favour of war. On 28 May 1941, Smuts was appointed a Field Marshal of the British Army, becoming the first South African to hold that rank. When the war ended, Smuts represented South Africa in San Francisco at the drafting of the United Nations Charter in May 1945. Just as he had done in 1919, Smuts urged the delegates to create a powerful international body to preserve peace; he was determined that, unlike the League of Nations, the UN would have teeth. Smuts also signed the Paris Peace Treaty, resolving the peace in Europe, thus becoming the only signatory of both the treaty ending the First World War, and that which ended the Second.
Pro-German and pro-Nazi attitudes
After the suppression of the abortive, pro-German Maritz Rebellion during the South African World War I campaign against German South West Africa in 1914, the South African rebel General Manie Maritz escaped to Spain. He returned in 1923, and continued working in the Union of South Africa as a German Spy for the Third Reich.
In 1896, the German Kaiser Kaiser Wilhelm had enraged Britain by sending congratulations to Boer republican leader Paul Kruger after Kruger's commandos captured a column of British South Africa Company soldiers engaged in an armed incursion and abortive insurrection, known historically as the Jameson Raid, into Boer territory. Germany was the primary supplier of weapons to the Boers during the subsequent Anglo–Boer war. Kaiser Wilhelm's government arranged for the two Boer Republics to purchase modern breech-loading Mauser rifles and millions of smokeless gunpowder cartridges. Germany's Ludwig Loewe company, later known as Deutsche Waffen-und Munitionfabriken, delivered 55,000 of these rifles to the Boers in 1896.
The early-1940s saw the pro-Nazi Ossewa Brandwag (OB) movement become half-a-million strong, including future prime minister John Vorster and Hendrik van den Bergh, the future head of police intelligence. The anti-semitic Boerenasie (Boer Nation) and other similar groups soon joined them. When the war ended, the OB was one of the anti-parliamentary groups absorbed into the National Party.
The South African Afrikaner Weerstandsbeweging or AWB (meaning Afrikaner Resistance Movement), a militant neo-Nazi, mainly Afrikaner white supremacist movement that arose in the 1970s, and was active until the mid-1990s, openly used a flag that closely resembled the swastika. In the early to mid-1990s, the AWB attempted unsuccessfully through various acts of public violence and intimidation to derail the country's transition to democracy. After the country's first multiracial democratic elections in 1994, a number of terrorist bomb blasts were linked to the AWB. On 11 March 1994, several hundred AWB members formed part of an armed right-wing force that invaded the nominally independent "homeland" territory of Bophuthatswana, in a failed attempt to prop up its unpopular, conservative leader Chief Lucas Mangope. The AWB leader Eugène Terre'Blanche was murdered by farm workers on 3 April 2010.
A majority of politically moderate Afrikaners were pragmatic and did not support the AWB's extremism.
Apartheid era (1948–1994)
Apartheid legislation
The segregationist policies of apartheid stemmed from colonial legislation introduced during the period of Dutch rule in the 17th century, which was continued and expanded upon during the British colonial era, and reached its apogee during the Boer-dominated Union of South Africa.
From 1948, successive National Party administrations formalised and extended the existing system of racial discrimination and denial of human rights into the legal system of apartheid, which lasted until 1991. A key act of legislation during this time was the Homeland Citizens Act of 1970. This act augmented the Native Land Act of 1913 through the establishment of so-called "homelands" or "reserves". It authorised the forced evictions of thousands of African people from urban centres in South Africa and South West Africa (now Namibia) to what became described colloquially as "Bantustans" or the "original homes", as they were officially referred to, of the black tribes of South Africa. The same legislation applied also to South West Africa over which South Africa had continued after World War I to exercise a disputed League of Nations mandate. Pro-apartheid South Africans attempted to justify the Bantustan policy by citing the British government's 1947 partition of India, which they claimed was a similar situation that did not arouse international condemnation.
Although many important events occurred during this period, apartheid remained the central pivot around which most of the historical issues of this period revolved, including violent conflict and the militarisation of South African society. By 1987, total military expenditure amounted to about 28% of the national budget.
In the aftermath of the 1976 Soweto uprising and the security clampdown that accompanied it, Joint Management Centres (JMCs) operating in at least 34 State-designated "high-risk" areas became the key element in a National Security Management System. The police and military who controlled the JMCs by the mid-1980s were endowed with influence in decision-making at every level, from the Cabinet down to local government.
UN embargo
On 16 December 1966, United Nations General Assembly Resolution 2202 A (XXI) identified apartheid as a "crime against humanity".
The Apartheid Convention, as it came to be known, was adopted by the General Assembly on 30 November 1973 with 91 member states voting in favour, four against (Portugal, South Africa, the United Kingdom and the United States) and 26 abstentions. The convention came into force on 18 July 1976. On 23 October 1984 the UN Security Council endorsed this formal determination. The convention declared that apartheid was both unlawful and criminal because it violated the Charter of the United Nations.
The General Assembly had already suspended South Africa from the UN organisation on 12 November 1974. On 4 November 1977, the Security Council imposed a mandatory arms embargo in terms of Resolution 181 calling upon all States to cease the sale and shipment of arms, ammunition and military vehicles to South Africa. The country would only be readmitted to the UN in 1994 following its transition to democracy. Apartheid South Africa reacted to the UN arms embargo by strengthening its military ties with Israel, and establishing its own arms manufacturing industry with the help of Israel. Four hundred M-113A1 armoured personnel carriers, and 106mm recoilless rifles manufactured in the United States were delivered to South Africa via Israel.
Extra-judicial killings
In the mid-1980s, police and army death squads conducted state-sponsored assassinations of dissidents and activists. By mid-1987 the Human Rights Commission knew of at least 140 political assassinations in the country, while about 200 people died at the hands of South African agents in neighbouring states. The exact numbers of all the victims may never be known. Strict censorship disallowed journalists from reporting, filming or photographing such incidents, while the government ran its own covert disinformation programme that provided distorted accounts of the extrajudicial killings. At the same time, State-sponsored vigilante groups carried out violent attacks on communities and community leaders associated with resistance to apartheid. The attacks were then falsely attributed by the government to "black-on-black" or factional violence within the communities.
The Truth and Reconciliation Commission (TRC) would later establish that a covert, informal network of former or still serving army and police operatives, frequently acting in conjunction with extreme right-wing elements, was involved in actions that could be construed as fomenting violence and which resulted in gross human rights violations, including random and targeted killings.
Between 1960–1994, according to statistics from the Truth and Reconciliation Commission, the Inkatha Freedom Party was responsible for 4,500 deaths, South African Police 2,700, and the ANC about 1,300.
In early 2002, a planned military coup by a white supremacist movement known as the Boeremag (Boer Force) was foiled by the South African police. Two dozen conspirators including senior South African Army officers were arrested on charges of treason and murder, after a bomb explosion in Soweto. The effectiveness of the police in foiling the planned coup strengthened public perceptions that the post-1994 democratic order was irreversible.
The TRC, at the conclusion of its mandate in 2004, handed over a list of 300 names of alleged perpetrators to the National Prosecuting Authority (NPA) for investigation and prosecution by the NPA's Priority Crimes Litigation Unit. Less than a handful of prosecutions were ever pursued.
Military operations in frontline states
South African security forces during the latter part of the apartheid era had a policy of destabilising neighbouring states, supporting opposition movements, conducting sabotage operations and attacking ANC bases and places of refuge for exiles in those states. These states, forming a regional alliance of southern African states, were named collectively as the Frontline States: Angola, Botswana, Lesotho, Mozambique, Swaziland, Tanzania, Zambia and, from 1980, Zimbabwe.
In early-November 1975, immediately after Portugal granted independence to its former African colony of Angola, civil war broke out between the rival UNITA and MPLA movements. In order to prevent UNITA's collapse and cement the rule of a friendly government, South Africa intervened on 23 October, sending between 1,500 and 2,000 troops from Namibia into southern Angola in order to fight the MPLA. In response to the South African intervention, Cuba sent 18,000 soldiers as part of a large-scale military intervention nicknamed Operation Carlota in support of the MPLA. Cuba had initially provided the MPLA with 230 military advisers prior to the South African intervention. The Cuban intervention was decisive in helping reverse SADF and UNITA advances and cement MPLA rule in Angola. More than a decade later 36,000 Cuban troops were deployed throughout the country helping providing support for MPLA's fight with UNITA. The civil war in Angola resulted in 550,000–1,250,000 deaths in total mostly from famine. Most of the deaths occurred between 1992 and 1993, after South African and Cuban involvement had ended.
Between 1975 and 1988, the SADF continued to stage massive conventional raids into Angola and Zambia to eliminate PLAN's forward operating bases across the border from Namibia as well as provide support for UNITA. A controversial bombing and airborne assault conducted by 200 South African paratroopers on 4 May 1978 at Cassinga in southern Angola, resulted in around 700 South West Africans being killed, including PLAN militants and a large number of women and children. Colonel Jan Breytenbach, the South African parachute battalion commander, claimed it was "recognised in Western military circles as the most successful airborne assault since World War II." The Angolan government described the target of the attack as a refugee camp. The United Nations Security Council on 6 May 1978 condemned South Africa for the attack. On 23 August 1981 South African troops again launched an incursion into Angola with collaboration and encouragement provided by the American Central Intelligence Agency (CIA). The Angolan army, in resisting what it perceived as a South African invasion, was supported by a combination of Cuban forces and PLAN and ANC guerrillas, all armed with weapons supplied by the Soviet Union. The apartheid-era South African military and political intelligence services, for their part, worked closely with American, British and West German secret services throughout the Cold War.
Both South Africa and Cuba claimed victory at the decisive battle of Cuito Cuanavale, which have been described as "the fiercest in Africa since World War II". However, the South African military had lost air superiority and its technological advantage, largely due to an international arms embargo against the country. South African involvement in Angola ended formally after the signing of a United Nations-brokered agreement known as the New York Accords between the governments of Angola, Cuba and South Africa, resulting in the withdrawal of all foreign troops from Angola and also South Africa's withdrawal from South West Africa (now Namibia), which the UN regarded as illegally occupied since 1966.
South Africa in the 1980s also provided logistical and other covert support to Resistência Nacional Moçambicana (RENAMO) rebels, in neighbouring Mozambique fighting the FRELIMO-run government during the Mozambique Civil War, and it launched cross-border raids into Lesotho, Swaziland and Botswana, killing or capturing a number of South African exiles.
Resistance to apartheid
Organised resistance to Afrikaner nationalism was not confined exclusively to activists of the oppressed, dark-skinned population. A movement known as the Torch Commando was formed in the 1950s, led by white war veterans who had fought fascism in Europe and North Africa during World War II, only to find fascism on the rise in South Africa when they returned home. With 250,000 paid-up members at the height of its existence, it was the largest white protest movement in the country's history. By 1952, the brief flame of mass-based white radicalism was extinguished, when the Torch Commando disbanded due to government legislation under the Suppression of Communism Act, 1950. Some members of the Torch Commando subsequently became leading figures in the armed wing of the banned African National Congress.
From the 1940s to the 1960s, anti-apartheid resistance within the country took the form mainly of passive resistance, influenced in part by the pacifist ideology of Mahatma Gandhi. After the March 1960 massacre of 69 peaceful demonstrators at Sharpeville, and the subsequent declaration of a state of emergency, and the banning of anti-apartheid parties including the African National Congress (ANC), the Pan-Africanist Congress (PAC), and the Communist Party of South Africa, the focus of national resistance turned to armed struggle and underground activity. The armed wing of the ANC, Umkhonto weSizwe (abbreviation MK, meaning Spear of the Nation) claimed moral legitimacy for the resort to violence on the grounds of necessary defence and just war. From the 1960s onwards until 1989, MK carried out numerous acts of sabotage and attacks on military and police personnel. The Truth and Reconciliation Commission noted in 2003 that, despite the ANC's stated policy of attacking only military and police targets, "the majority of casualties of MK operations were civilians."
The national liberation movement was divided in the early 1960s when an "Africanist" faction within the ANC objected to an alliance between the ANC and the Communist Party of South Africa. Leaders of the Communist Party of South Africa were mostly white. The Africanists broke away from the ANC to form the Pan-Africanist Congress and its military wing named Poqo, which became active mainly in the Cape provinces. During the early-1990s, Poqo was renamed Azanian People's Liberation Army (APLA). Its underground cells conducted armed robberies to raise funds and obtain weapons and vehicles. Civilians were killed or injured in many of these robberies. In 1993, attacks on white civilian targets in public places increased. APLA denied the attacks were racist in character, claiming that the attacks were directed against the apartheid government as all whites, according to the PAC, were complicit in the policy of apartheid. An attack on a Christian church in Cape Town in 1993, left eleven people dead and 58 injured.
Hundreds of students and others who fled to neighbouring countries, especially Botswana, to avoid arrest after the Soweto uprising of 16 June 1976, provided a fertile recruiting ground for the military wings of both the ANC and PAC. The uprising had been precipitated by Government legislation forcing African students to accept Afrikaans as the official medium for tuition, with support from the wider Black Consciousness Movement. The uprising spread throughout the country. By the time it was finally quelled, hundreds of protesters had been shot dead with many more wounded or arrested by police.
A non-racial United Democratic Front (UDF) coalition of about 400 civic, church, student, trade union and other organisations emerged in 1983. At its peak in 1987, the UDF had some 700 affiliates and about 3,000,000 members. It pursued a non-violent strategy known as "ungovernability" including rent boycotts, student protests, and strike campaigns. A strong relationship existed between the African National Congress (ANC) and the UDF, based on the shared mission statement of the Freedom Charter. Following restrictions placed on its activities, the UDF was replaced in 1988 by the Mass Democratic Movement, a loose and amorphous alliance of anti-apartheid groups that had no permanent structure, making it difficult for the government to place a ban on its activities.
A total of 130 political prisoners were hanged on the gallows of Pretoria Central Prison between 1960 and 1990. The prisoners were mainly members of the Pan Africanist Congress and United Democratic Front.
Post-apartheid period (1994–present)
The dissolution of the Soviet Union in the late-1980s meant the African National Congress (ANC) in alliance with the South African Communist Party, could no longer depend on the Soviet Union for weaponry and political support. It also meant the apartheid government could no longer link apartheid and its purported legitimacy to the protection of Christian values and civilisation in the face of the rooi gevaar, meaning "red danger" or the threat of communism. Both sides were forced to the negotiating table, with the result that in June 1991, all apartheid laws were finally rescinded- opening the way for the country's first multiracial democratic elections three years later. As the culmination of mounting local and international opposition to apartheid in the 1980s, including the armed struggle, widespread civil unrest, economic and cultural sanctions by the international community, and pressure from the anti-apartheid movement around the world, State President F. W. de Klerk announced the lifting of the ban on the African National Congress, the Pan Africanist Congress and the South African Communist Party, as well as the release of political prisoner Nelson Mandela on 2 February 1990, after twenty-seven years in prison. In a referendum held on 17 March 1992, the white electorate voted 68% in favour of democracy.
After lengthy negotiations under the auspices of the Convention for a Democratic South Africa (CODESA), a draft constitution was published on 26 July 1993, containing concessions towards all sides: a federal system of regional legislatures, equal voting-rights regardless of race, and a bicameral legislature.
From 26–29 April 1994, the South African population voted in the first universal suffrage general elections. The African National Congress won, well ahead of the governing National Party and the Inkatha Freedom Party. The Democratic Party and Pan Africanist Congress, among others, formed a parliamentary opposition in the country's first non-racial parliament. Nelson Mandela was elected as President on 9 May 1994 and formed a Government of National Unity, consisting of the ANC, the National Party and Inkatha. On 10 May 1994 Mandela was inaugurated as South Africa's new President in Pretoria with Thabo Mbeki and F. W. De Klerk as his vice-presidents. The Government of National Unity lapsed at the end of the first parliament sitting in 1999, with the ANC becoming the sole party in power while maintaining a strategic alliance with the Congress of South African Trade Unions (COSATU) and the South African Communist Party. After considerable debate, and following submissions from advocacy groups, individuals and ordinary citizens, the Parliament enacted a new Constitution and Bill of Rights in 1996. The death penalty was abolished, land reform and redistribution policies were introduced, and equitable labour laws legislated.
The ANC had risen to power on the strength of a socialist agenda embodied in a Freedom Charter, which was intended to form the basis of ANC social, economic and political policies. The Charter decreed that "the national wealth of our country, the heritage of South Africans, shall be restored to the people; the mineral wealth beneath the soil, the banks and monopoly industry shall be transferred to the ownership of the people". ANC icon Nelson Mandela, asserted in a statement released on 25 January 1990: "The nationalisation of the mines, banks and monopoly industries is the policy of the ANC, and a change or modification of our views in this regard is inconceivable."
Following the ANC's electoral victory in 1994, the eradication of mass poverty through nationalisation was never implemented. The ANC-led government, in a historic reversal of policy, adopted neoliberalism instead. A wealth tax on the super-rich to fund developmental projects was set aside, while domestic and international corporations, enriched by apartheid, were excused from any financial reparations. Large corporations were allowed to shift their main listings abroad. According to Solomon Johannes Terreblanche, a South African academic economist, the government's concessions to big business represented "treacherous decisions that [will] haunt South Africa for generations to come".
Emigration
The immediate post-apartheid period was marked by an exodus of skilled, white South Africans amid crime related safety concerns. The South African Institute of Race Relations estimated in 2008 that 800,000 or more white people had emigrated since 1995, out of the approximately 4,000,000 who were in South Africa when apartheid formally ended the year before. Large white South African diasporas, both English- and Afrikaans-speaking, sprouted in Australia, New Zealand, North America, and especially in the UK, to which around 550,000 South Africans emigrated. As of 2021, tens of thousands of white South Africans continue to emigrate each year.
Financial burdens
The apartheid government had declared a moratorium on foreign debt repayments in the mid-1980s, when it declared a state of emergency in the face of escalating civil unrest. With the formal end of apartheid in 1994, the new democratic government was saddled with an onerous foreign debt amounting to R86,700,000,000 (US$14,000,000,000 at then current exchange rates) accrued by the former apartheid regime. The cash-strapped post-apartheid government was obliged to repay this debt or else face a credit downgrading by foreign financial institutions. The debt was finally settled in September 2001.
A further financial burden was imposed on the new post-apartheid government through its obligation to provide antiretroviral (ARV) treatment to impoverished victims of the HIV/AIDS epidemic sweeping the country. South Africa had the highest prevalence of HIV/AIDS compared to any other country in the world, with 5,600,000 people afflicted by the disease and 270,000 HIV-related deaths were recorded in 2011. By that time, more than 2,000,000 children were orphaned due to the epidemic. The provision of ARV treatment resulted in 100,000 fewer AIDS-related deaths in 2011 than in 2005.
Labour relations
Migrant labour remained a fundamental aspect of the South African mining industry, which employed half a million mostly black miners. Labour unrest in the industry resulted in a massacre in mid-August 2012, when anti-riot police shot dead 34 striking miners and wounded many more in what is known as the Marikana massacre. The incident was widely criticised by the public, civil society organisations and religious leaders. The migrant labour system was identified as a primary cause of the unrest. Multi-national mining corporations including Anglo-American Corporation, Lonmin, and Anglo Platinum, were accused of failing to address the enduring legacies of apartheid.
Poverty
By 2014, around 47% of (mostly black) South Africans continued to live in poverty, making it one of the most unequal countries in the world. Widespread dissatisfaction with the slow pace of socio-economic transformation, government incompetence and maladministration, and other public grievances in the post-apartheid era, precipitated many violent protest demonstrations. In 2007, less than half the protests were associated with some form of violence, compared with 2014, when almost 80% of protests involved violence on the part of the participants or the authorities. The slow pace of transformation also fomented tensions within the tripartite alliance between the ANC, the Communist Party and the Congress of South African Trade Unions.
Corruption
During the administration of President Jacob Zuma corruption in South Africa had also become a growing problem. Notable corruption related scandals during this period included incidents of widespread state capture often involving allegations against the Gupta family. These also involved corruption related financial difficulties at some state owned enterprises such as Eskom and South African Airways that had a notable negative economic impact on the country's finances. Other corruption related scandals to emerge during this period included the collapse of VBS Mutual Bank and Bosasa. The Zondo Commission of Inquiry was appointed during the Presidency of Cyril Ramaphosa to investigate allegations of state capture related corruption.
Xenophobia
The post-apartheid period has been marked by numerous outbreaks of xenophobic attacks against foreign migrants and asylum seekers from various conflict zones in Africa. An academic study conducted in 2006, found that South Africans showed levels of xenophobia greater than anywhere else in the world. The United Nations High Commissioner for Refugees (UNHCR) found that competition over jobs, business opportunities, public services and housing gave rise to tension among refugees, asylum seekers, migrants and host communities, identified as a main cause of the xenophobic violence. South Africa received more than 207,000 individual asylum applications in 2008 and a further 222,300 in 2009, representing nearly a four-fold rise in both years over the levels seen in 2007. These refugees and asylum seekers originated mainly from Zimbabwe, Burundi, Democratic Republic of the Congo, Rwanda, Eritrea, Ethiopia and Somalia.
2021 civil unrest
Civil unrest occurred in South Africa's KwaZulu-Natal and Gauteng provinces in July 2021, sparked by the imprisonment of former President Jacob Zuma for contempt of court, after he declined to testify at the Zondo Commission, an inquiry into allegations of corruption during his term as president from 2009 to 2018. Protests against the incarceration triggered wider rioting and looting, further exacerbated by job layoffs and economic woes worsened by the COVID-19 pandemic.
The Economist called it the worst violence that South Africa had experienced since the end of Apartheid.
Police and military authorities were mobilised to quell the unrest. By mid-July, the South African National Defense Forces had deployed approximately 25,000 military personnel. As of 18 July, over 3,400 people had been arrested, while as of 22 July, 337 people had died in connection with the unrest.
The July 2021 unrest coincided with the Cape Town taxi conflict and Transnet ransomware attack leading to unproven speculation that they might have been connected.
Post-apartheid heads of state
Under the post-apartheid Constitution the president is head of both state and government. The president is elected by the National Assembly and serves a term that expires at the next general election. A president may serve a maximum of two terms. In the event of a vacancy the Deputy President serves as Acting President.
See also
History of Africa
Scramble for Africa
History of Cape Colony
History of Johannesburg
History of the Northern Cape
History of South African wine
List of presidents of South Africa
List of prime ministers of South Africa
List of heads of state of South Africa
List of South Africa-related topics
Military history of South Africa
Politics of South Africa
Timeline of liberal parties in South Africa
Years in South Africa
History of cities in South Africa:
Cape Town history and timeline
Durban history and timeline
Johannesburg history and timeline
Pietermaritzburg history and timeline
Port Elizabeth history and timeline
Pretoria history and timeline
References
Further reading
General
Beinart, William. Twentieth-Century South Africa. Oxford University Press. 2001.
Beck, Roger S. History of South Africa, Greenwood Press, Westport CT: 2000
Blignaut, Charl. "Untold History with a Historiography: A Review of Scholarship on Afrikaner Women in South African History." South African Historical Journal 65.4 (2013): 596–617.
Bunting, Brian. Rise of the South African Reich First published by Penguin Africa Library 1964, revised 1969.
Christopher, A. J. The Atlas of Changing South Africa. 2000. 216 pages. .
Deegan, Heather. The Politics of the New South Africa. 2000. 256 pages. .
Elbourne, Elizabeth. Blood Ground: Colonialism, Missions, and the Contest for Christianity in the Cape Colony and Britain, 1799–1853. McGill-Queen's University Press. December 2002. 560 pages. .
Hamilton, Carolyn and Bernard K. Mbenga, eds. The Cambridge History of South Africa: Volume 1: From Early Times to 1885 (2009) excerpt
Hetherington, Penelope. "Women in South Africa: the historiography in English." The International Journal of African Historical Studies 26#2 (1993): 241–269.
Johnson, R.W. South Africa's Brave New World: The Beloved Country Since the End of Apartheid (Overlook Press; 2011) 702 pages; a history since 1994
Joyce, Peter. The Making of a Nation South Africa's Road to Freedom, Zebra Press, 2004,
Le Cordeur, Basil Alexander. The War of the Axe, 1847: Correspondence between the governor of the Cape Colony, Sir Henry Pottinger, and the commander of the British forces at the Cape, Sir George Berkeley, and others. Brenthurst Press. 1981. 287 pages. .
Mabin, Alan. Recession and its aftermath: The Cape Colony in the eighteen eighties. University of the Witwatersrand, African Studies Institute. 1983. 27 pages.
Meiring, Hannes. Early Johannesburg, Its Buildings and People, Human & Rousseau, 1986, 143 pages,
Mitchell, Laura. Belongings: Property, Family, and Identity in Colonial South Africa: An Exploration of Frontiers, 1725-c. 1830. Columbia University Press, 2008. Gutenberg-e.org
Pakenham, Thomas. The Boer War, Weidenfeld and Nicolson, London 1979,
Rosenthal, Eric. Gold! Gold! Gold! The Johannesburg Gold Rush, AD. Donker, 1970,
Ross, Robert, and David Anderson. Status and Respectability in the Cape Colony, 1750–1870 : A Tragedy of Manners. Cambridge University Press. 1999. 220 pages. .
Ross, Robert, Anne Kelk Mager and Bill Nasson, eds. The Cambridge History of South Africa: Volume 2 since 1885 (2011) excerpt
Thompson, Leonard. A History of South Africa, Third Edition. Yale University Press. 2001. 384 pages. .
Tomlinson, Richard, et al. Emerging Johannesburg: Perspectives on the Postapartheid City. 2003. 336 pages. .
Welsh, Frank. South Africa: A Narrative History. Kodansha America. 1 February 1999. 606 pages. .
Worden, Nigel. Making of Modern South Africa: Conquest, Segregation and Apartheid. 2000. 194 pages. .
VOC period
Barend-van Haeften, Marijke; Paasman, Bert: De Kaap: Goede Hoop halverwege Indië. Bloemlezing van Kaapteksten uit de Compagnietijd. (Hilversum: Verloren, 2003)
Biewenga, A.: De Kaap de Goede Hoop: Een Nederlandse Vestigingskolonie, 1680–1730. (Amsterdam: Promotheus and Bert Bakker, 1999)
Botha, Colin Graham: The French Refugees at the Cape. (1919; reprint, Cape Town: C. Struik, 1970)
Bryer, Lynne; Theron, Francois: The Huguenot Heritage: The Story of the Huguenots at the Cape. (Diep River, Chameleon Press, 1987)
Coertzen, Pieter: Die Hugenote van Suid Afrika, 1688–1988. (Cape Town: Tafelberg Publishers Limited, 1988)
Delmas, Adrien, 'The Role of Writing in the First Steps of the Colony: A Short Enquiry in the Journal of Jan van Riebeeck, 1652–1662', in Contingent Lives: Social Identity and Material Culture in the VOC World, ed. Nigel Worden (Cape Town: Royal Netherlands Embassy, 2007)
Elphick, Richard; Giliomee, Hermann (eds.): The Shaping of South African Society, 1652–1840. (Wesleyan University Press, 1989, )
Fourie, J.; Boshoff, W. (2008). ‘Explaining the Ship Traffic Fluctuations in the Early Cape Settlement: 1652–1793’, South African Journal of Economic History, 23 (2008), pp. 1–27.
Fourie, J.; Boshoff, W. (2010). ‘The significance of the Cape trade route to economic activity in the Cape Colony: a medium-term business cycle analysis’, European Review of Economic History, 14 (2010), pp. 469–503.
Fourie, J. (2014). ‘The quantitative Cape: A review of the new historiography of the Dutch Cape Colony’, South African Historical Journal 66.1, 2014, pp. 142–168.
Franken, J. L. N.: Die Hugenote aan die Kaap. (Pretoria: Die Staatsdrukker, 1978)
Gerstner, Jonathan Neil: The Thousand Generation Covenant: Dutch Reformed Covenant Theology and Group Identity in Colonial South Africa, 1652–1814. (Leiden: E.J. Brill, 1991)
Godée Molsbergen, E.C.: Reizen in Zuid-Afrika in de Hollandse tijd: eerste deel, Tochten naar het noorden, 1652–1686. ('s-Gravenhage: Martinus Nijhoff, 1976)
Godée Molsbergen, E.C.: Reizen in Zuid-Afrika in de Hollandse tijd: tweede deel, Tochten naar het noorden, 1686–1806. ('s-Gravenhage: Martinus Nijhoff, 1976)
Godée Molsbergen, E.C.: Jan van Riebeeck en zijn tijd. Een stuk zeventiende-eeuws Oost-Indië [Jan van Riebeeck and his times: A piece of seventeenth-century East Indies]. (Amsterdam: P. N. Van Kampen en Zoon, 1937)
Groenewald, Gerald: Een Dienstig Inwoonder: Entrepreneurs, Social Capital and Identity in Cape Town, c. 1720–1750. South African Historical Journal, 59, 1(2007), pp. 126–152
Guelke, Leonard (1976). "Frontier Settlement in Early Dutch South Africa,". (Annals of the Association of American Geographers 66, no. 1 (March 1976): 25–42)
Huigen, Siegfried: Knowledge and Colonialism: Eighteenth-Century Travellers in South Africa. (Leiden: Brill, 2009, pp. xii + 273 pp)
Hunt, John: Dutch South Africa: Early Settlers at the Cape, 1652–1708. (Leicester, UK: Matador, 2005, )
Johnson, David: Imagining the Cape Colony: History, Literature and the South African Nation. (Edinburgh: Edinburgh University Press, 2012)
Lucas, Gavin: An Archaeology of Colonial Identity: Power and Material Culture in the Dwars Valley, South Africa. (New York: Springer, 2006)
Marais, J.S.: The Cape Coloured People, 1652–1937. (1939; reprint, Johannesburg: Witwatersrand University Press, 1968)
Mitchell, L.J.: Belongings: Property and Identity in Colonial South Africa, an Exploration of Frontiers, 1725–c. 1830. (New York: Columbia University Press, 2009, pp. xv + 232)
Newton-King, Susan: Masters and Servants on the Cape Eastern Frontier, 1760–1803. (Cambridge: Cambridge University Press, 1999)
Raven-Hart, Rowland (ed.): Cape Good Hope, 1652–1702: The First Fifty Years of Dutch Colonisation as Seen by Callers [2 vols]. (Cape Town: A.A.Balkema, 1971)
Schoeman, Karel: Early Slavery at the Cape of Good Hope, 1652–1717. (Pretoria: Protea Book House, 2007)
Schoeman, Karel: Portrait of a Slave Society: The Cape of Good Hope, 1717–1795. (Pretoria: Protea Boekhuis, 2012)
Shell, Robert: Children of Bondage: A Social History of the Slave Society at the Cape of Good Hope, 1652–1838. (Hanover, Conn.; Wesleya University Press, 1994)
Singh, Daleep: From Dutch South Africa to Republic of South Africa, 1652–1994. The Story of Three and a Half Centuries of Imperialism. (New Delhi: Allied Publishers, 2010)
Sleigh, Dan: Die Buiteposte: VOC-buiteposte onder Kaapse Bestuur, 1652–1795. (Pretoria: HAUM, 1993)
Stapleton, Timothy J.: A Military History of South Africa: From the Dutch-Khoi Wars to the End of Apartheid. (Santa Barbara, CA.: Praeger, 2010)
Terreblanche, Sampie: A History of Inequality in South Africa, 1652–2002. (Pietermaritzburg: University of Natal Press, 2002)
Theal, George McCall: History of South Africa under the Administration of the Dutch East India Company, 1652 to 1795. Nabu Press, 2010, 474pp
Theal, George McCall: History of the Boers in South Africa; Or, the Wanderings and Wars of the Emigrant Farmers from Their Leaving the Cape Colony to the Acknowledgment of Their Independence by Great Britain. (Greenwood Press, 1970, )
Twidle, Hedley (2013). Writing the Company: From VOC Daghregister to Sleigh's Eilande, South African Historical Journal 65(1) (2013): 125–52.
Valentijn, Francois: Description of the Cape of Good Hope with Matters Concerning It. (Amsterdam, 1726). [Edited and annotated by Prof. P. Serton, Maj. B. Raven-Hart, Dr. W. J. de Kock.]
Van der Merwe, P. J. : The Migrant Farmer in the History of the Cape Colony, 1657–1842. Translated from the Dutch by Roger B. Beck. (Athens, Ohio: Ohio University Press, 1995)
Van Duin, Pieter; Ross, Robert: The Economy of the Cape Colony in the Eighteenth Century. (Leiden: Centre for the History of European Expansion, 1987, pp. viii + 166)
Van Riebeeck, Jan: 1652–1662, Daghregister gehouden by den oppercoopman Jan Anthonisz van Riebeeck. In Bosman, D.B. & B. Thorn (eds). Daghregister Gehouden by den Oppercoopman Jan Anthonisz van Riebeeck, 3 vols. [Treated as a sequence of eleven manuscripts]. (Cape Town: Balkema, 1952–1957)
Verstegen, M.: De Indische Zeeherberg, De stichting van Zuid-Afrika door de VOC. (Zaltbommel, 2001)
Viljoen, Russel: Jan Paerl: A Khoikhoi in Cape Colonial Society, 1761–1851. (Boston, MA: Brill, 2006, pp. xviii + 213)
Viljoen, Russel S. (1995). "Disease and Society: VOC Cape Town, Its People, and the Smallpox Epidemics of 1713, 1755, and 1767," Kleio 27
Welch, Sidney: Portuguese and Dutch in South Africa, 1641–1806. (Cape Town: Juta Press, 1951)
External links
South African History Archive
Democratic Development in South Africa from the Dean Peter Krogh Foreign Affairs Digital Archives
South African History Online
The History of South Africa Brand South Africa. Accessed 24 November 2017.
South Africa Government Online. Accessed 20 February 2005.
Dr Cyril Hromník on research into ancient history of Africa – an article written by Maré Mouton.
Bearer of an Ideal – a public-release document of the Afrikanerbond (formerly Afrikaner Broederbond): think-tank which influenced policies of separate development in South Africa
Full text of the UN convention
South Africa, 10 years later from US National Public Radio
(Directory of South African archival and memory institutions and organisations)
Articles containing video clips |
null | null | Metcalfe's law | eng_Latn | Metcalfe's law states that the value of a telecommunications network is proportional to the square of the number of connected users
of the system (n2). First formulated in this form by George Gilder in 1993, and attributed to Robert Metcalfe in regard to Ethernet, Metcalfe's law was originally presented, c. 1980, not in terms of users, but rather of "compatible communicating devices" (e.g., fax machines, telephones). Only later with the globalization of the Internet did this law carry over to users and networks as its original intent was to describe Ethernet connections.
Network effects
Metcalfe's law characterizes many of the network effects of communication technologies and networks such as the Internet, social networking and the World Wide Web. Former Chairman of the U.S. Federal Communications Commission Reed Hundt said that this law gives the most understanding to the workings of the Internet. Metcalfe's Law is related to the fact that the number of unique possible connections in a network of nodes can be expressed mathematically as the triangular number , which is asymptotically proportional to .
The law has often been illustrated using the example of fax machines: a single fax machine is useless, but the value of every fax machine increases with the total number of fax machines in the network, because the total number of people with whom each user may send and receive documents increases. Likewise, in social networks, the greater the number of users with the service, the more valuable the service becomes to the community.
History and Derivation
Metcalfe’s law was conceived in 1983 in a presentation to the 3Com sales force. It stated V would be proportional to the total number of possible connections, or approximately n-squared.
The original incarnation was careful to delineate between a linear cost (Cn), non-linear growth, n2, and a non-constant proportionality factor A “Affinity.” The breakeven point where costs are recouped is given by
At some size, the right-hand side of the equation V “Value” exceeds the cost, and A describes the relationship between size and net value added. For large n, net network value is then
Metcalfe properly dimensioned A as “value per user”. Affinity is also a function of network size, and Metcalfe correctly asserted that A must decline as n grows large. In a 2006 interview, Metcalfe stated
“There may be dis-economies of network scale that eventually drive values down with increasing size. So, if V=A*n2, it could be that A (for “affinity,” value per connection) is also a function of n and heads down after some network size, overwhelming n2.”
Growth of n
Network size, and hence value, does not grow unbounded but is constrained by practical limitations such as infrastructure, access to technology, and bounded rationality such as Dunbar's number. It is almost always the case that user growth n reaches a saturation point. With technologies, substitutes, competitors and technical obsolescence constrain growth of n. Growth of n is typically assumed to follow a sigmoid function such as a logistic curve or Gompertz curve.
Density
A is also governed by the connectivity or density of the network topology. In an undirected network, every edge connects two nodes such that there are 2m nodes per edge. The proportion of nodes in actual contact are given by .
The maximum possible number of edges in a simple network (i.e. one with no multi-edges or self-edges) is .
Therefore the density ρ of a network is the faction of those edges that are actually present is
which for large networks is approximated by .
Limitations
Metcalfe’s law assumes that the value of each node is of equal benefit. If this is not the case, for example because one fax machine serves 50 workers in a company, the second fax machine serves half of that, the third one third, and so on, then the relative value of an additional connection decreases. Likewise, in social networks, if users that join later use the network less than early adopters, then the benefit of each additional user may lessen, making the overall network less efficient if costs per users are fixed.
Modified models
Within the context of social networks, many, including Metcalfe himself, have proposed modified models in which the value of the network grows as rather than . Reed and Andrew Odlyzko have sought out possible relationships to Metcalfe's Law in terms of describing the relationship of a network and one can read about how those are related. Tongia and Wilson also examine the related question of the costs to those excluded.
Validation in data
Despite many arguments about Metcalfe' law, no real data based evidence for or against was available for more than 30 years. Only in July 2013, Dutch researchers managed to analyze European Internet usage patterns over a long enough time and found proportionality for small values of and proportionality for large values of . A few months later, Metcalfe himself provided further proof, as he used Facebook's data over the past 10 years to show a good fit for Metcalfe's law (the model is ).
In 2015, Zhang, Liu and Xu parameterized the Metcalfe function in data from Tencent and Facebook. Their work showed that Metcalfe's law held for both, despite differences in audience between the two sites (Facebook serving a worldwide audience and Tencent serving only Chinese users). The functions for the two sites were and respectively.
In a working paper, Peterson linked time-value-of-money concepts to Metcalfe value using Bitcoin and Facebook as numerical examples of the proof and in 2018 applied Metcalfe's law to Bitcoin, showing that over 70% of variance in Bitcoin value was explained by applying Metcalfe's law to increases in Bitcoin network size.
See also
Anti-rival good
Beckstrom's law
List of eponymous laws
Matching (graph theory)
Matthew effect
Pareto principle
Reed's law
Sarnoff's law
References
Further reading
.
External links
A Group Is Its Own Worst Enemy. Clay Shirky's keynote speech on Social Software at the O'Reilly Emerging Technology conference, Santa Clara, April 24, 2003. The fourth of his "Four Things to Design For" is: "And, finally, you have to find a way to spare the group from scale. Scale alone kills conversations, because conversations require dense two-way conversations. In conversational contexts, Metcalfe's law is a drag."
Computer architecture statements
Economics laws
Eponyms
Information theory |
null | null | Alternation of generations | eng_Latn | Alternation of generations (also known as metagenesis or heterogenesis) is the type of life cycle that occurs in those plants and algae in the Archaeplastida and the Heterokontophyta that have distinct haploid sexual and diploid asexual stages. In these groups, a multicellular haploid gametophyte with n chromosomes alternates with a multicellular diploid sporophyte with 2n chromosomes, made up of n pairs. A mature sporophyte produces haploid spores by meiosis, a process which reduces the number of chromosomes to half, from 2n to n.
The haploid spores germinate and grow into a haploid gametophyte. At maturity, the gametophyte produces gametes by mitosis, which does not alter the number of chromosomes. Two gametes (originating from different organisms of the same species or from the same organism) fuse to produce a diploid zygote, which develops into a diploid sporophyte. This cycle, from gametophyte to sporophyte (or equally from sporophyte to gametophyte), is the way in which all land plants and many algae undergo sexual reproduction.
The relationship between the sporophyte and gametophyte varies among different groups of plants. In those algae which have alternation of generations, the sporophyte and gametophyte are separate independent organisms, which may or may not have a similar appearance. In liverworts, mosses and hornworts, the sporophyte is less well developed than the gametophyte and is largely dependent on it. Although moss and hornwort sporophytes can photosynthesise, they require additional photosynthate from the gametophyte to sustain growth and spore development and depend on it for supply of water, mineral nutrients and nitrogen. By contrast, in all modern vascular plants the gametophyte is less well developed than the sporophyte, although their Devonian ancestors had gametophytes and sporophytes of approximately equivalent complexity. In ferns the gametophyte is a small flattened autotrophic prothallus on which the young sporophyte is briefly dependent for its nutrition. In flowering plants, the reduction of the gametophyte is much more extreme; it consists of just a few cells which grow entirely inside the sporophyte.
Animals develop differently. They produce haploid gametes. No haploid spores capable of dividing are produced, so generally there is no multicellular haploid phase. (Some insects have a sex-determining system whereby haploid males are produced from unfertilized eggs; however females produced from fertilized eggs are diploid.)
Life cycles of plants and algae with alternating haploid and diploid multicellular stages are referred to as diplohaplontic (the equivalent terms haplodiplontic, diplobiontic and dibiontic are also in use, as is describing such an organism as having a diphasic ontogeny). Life cycles, such as those of animals, in which there is only a diploid multicellular stage are referred to as diplontic. Life cycles in which there is only a haploid multicellular stage are referred to as haplontic.
Definition
Alternation of generations is defined as the alternation of multicellular diploid and haploid forms in the organism's life cycle, regardless of whether these forms are free-living. In some species, such as the alga Ulva lactuca, the diploid and haploid forms are indeed both free-living independent organisms, essentially identical in appearance and therefore said to be isomorphic. The free-swimming, haploid gametes form a diploid zygote which germinates into a multicellular diploid sporophyte. The sporophyte produces free-swimming haploid spores by meiosis that germinate into haploid gametophytes.
However, in some other groups, either the sporophyte or the gametophyte is very much reduced and is incapable of free living. For example, in all bryophytes the gametophyte generation is dominant and the sporophyte is dependent on it. By contrast, in all modern vascular land plants the gametophytes are strongly reduced, although the fossil evidence indicates that they were derived from isomorphic ancestors. In seed plants, the female gametophyte develops totally within the sporophyte, which protects and nurtures it and the embryonic sporophyte that it produces. The pollen grains, which are the male gametophytes, are reduced to only a few cells (just three cells in many cases). Here the notion of two generations is less obvious; as Bateman & Dimichele say "[s]porophyte and gametophyte effectively function as a single organism". The alternative term 'alternation of phases' may then be more appropriate.
History
Debates about alternation of generations in the early twentieth century can be confusing because various ways of classifying "generations" co-exist (sexual vs. asexual, gametophyte vs. sporophyte, haploid vs. diploid, etc.).
Initially, Chamisso and Steenstrup described the succession of differently organized generations (sexual and asexual) in animals as "alternation of generations", while studying the development of tunicates, cnidarians and trematode animals. This phenomenon is also known as heterogamy. Presently, the term "alternation of generations" is almost exclusively associated with the life cycles of plants, specifically with the alternation of haploid gametophytes and diploid sporophytes.
Wilhelm Hofmeister demonstrated the morphological alternation of generations in plants, between a spore-bearing generation (sporophyte) and a gamete-bearing generation (gametophyte). By that time, a debate emerged focusing on the origin of the asexual generation of land plants (i.e., the sporophyte) and is conventionally characterized as a conflict between theories of antithetic (Čelakovský, 1874) and homologous (Pringsheim, 1876) alternation of generations. Čelakovský coined the words sporophyte and gametophyte.
Eduard Strasburger (1874) discovered the alternation between diploid and haploid nuclear phases, also called cytological alternation of nuclear phases. Although most often coinciding, morphological alternation and nuclear phases alternation are sometimes independent of one another, e.g., in many red algae, the same nuclear phase may correspond to two diverse morphological generations. In some ferns which lost sexual reproduction, there is no change in nuclear phase, but the alternation of generations is maintained.
Alternation of generations in plants
Fundamental elements
The diagram above shows the fundamental elements of the alternation of generations in plants. The many variations found in different groups of plants are described by use of these concepts later in the article. Starting from the right of the diagram, the processes involved are as follows:
Two single-celled haploid gametes, each containing n unpaired chromosomes, fuse to form a single-celled diploid zygote, which now contains n pairs of chromosomes, i.e. 2n chromosomes in total.
The single-celled diploid zygote germinates, dividing by the normal process (mitosis), which maintains the number of chromosomes at 2n. The result is a multi-cellular diploid organism, called the sporophyte (because at maturity it produces spores).
When it reaches maturity, the sporophyte produces one or more sporangia (singular: sporangium) which are the organs that produce diploid spore mother cells (sporocytes). These divide by a special process (meiosis) that reduces the number of chromosomes by a half. This initially results in four single-celled haploid spores, each containing n unpaired chromosomes.
The single-celled haploid spore germinates, dividing by the normal process (mitosis), which maintains the number of chromosomes at n. The result is a multi-cellular haploid organism, called the gametophyte (because it produces gametes at maturity).
When it reaches maturity, the gametophyte produces one or more gametangia (singular: gametangium) which are the organs that produce haploid gametes. At least one kind of gamete possesses some mechanism for reaching another gamete in order to fuse with it.
The 'alternation of generations' in the life cycle is thus between a diploid (2n) generation of sporophytes and a haploid (n) generation of gametophytes.
The situation is quite different from that in animals, where the fundamental process is that a diploid (2n) individual produces haploid (n) gametes by meiosis. In animals, spores (i.e. haploid cells which are able to undergo mitosis) are not produced, so there is no asexual multi-cellular generation. Some insects have haploid males that develop from unfertilized eggs, but the females are all diploid.
Variations
The diagram shown above is a good representation of the life cycle of some multi-cellular algae (e.g. the genus Cladophora) which have sporophytes and gametophytes of almost identical appearance and which do not have different kinds of spores or gametes.
However, there are many possible variations on the fundamental elements of a life cycle which has alternation of generations. Each variation may occur separately or in combination, resulting in a bewildering variety of life cycles. The terms used by botanists in describing these life cycles can be equally bewildering. As Bateman and Dimichele say "[...] the alternation of generations has become a terminological morass; often, one term represents several concepts or one concept is represented by several terms."
Possible variations are:
Relative importance of the sporophyte and the gametophyte.
Equal (homomorphy or isomorphy).Filamentous algae of the genus Cladophora, which are predominantly found in fresh water, have diploid sporophytes and haploid gametophytes which are externally indistinguishable. No living land plant has equally dominant sporophytes and gametophytes, although some theories of the evolution of alternation of generations suggest that ancestral land plants did.
Unequal (heteromorphy or anisomorphy).
Dominant gametophyte (gametophytic).In liverworts, mosses and hornworts, the dominant form is the haploid gametophyte. The diploid sporophyte is not capable of an independent existence, gaining most of its nutrition from the parent gametophyte, and having no chlorophyll when mature.
Dominant sporophyte (sporophytic).In ferns, both the sporophyte and the gametophyte are capable of living independently, but the dominant form is the diploid sporophyte. The haploid gametophyte is much smaller and simpler in structure. In seed plants, the gametophyte is even more reduced (at the minimum to only three cells), gaining all its nutrition from the sporophyte. The extreme reduction in the size of the gametophyte and its retention within the sporophyte means that when applied to seed plants the term 'alternation of generations' is somewhat misleading: "[s]porophyte and gametophyte effectively function as a single organism". Some authors have preferred the term 'alternation of phases'.
Differentiation of the gametes.
Both gametes the same (isogamy).Like other species of Cladophora, C. callicoma has flagellated gametes which are identical in appearance and ability to move.
Gametes of two distinct sizes (anisogamy).
Both of similar motility.Species of Ulva, the sea lettuce, have gametes which all have two flagella and so are motile. However they are of two sizes: larger 'female' gametes and smaller 'male' gametes.
One large and sessile, one small and motile (oogamy). The larger sessile megagametes are eggs (ova), and smaller motile microgametes are sperm (spermatozoa, spermatozoids). The degree of motility of the sperm may be very limited (as in the case of flowering plants) but all are able to move towards the sessile eggs. When (as is almost always the case) the sperm and eggs are produced in different kinds of gametangia, the sperm-producing ones are called antheridia (singular antheridium) and the egg-producing ones archegonia (singular archegonium).
Antheridia and archegonia occur on the same gametophyte, which is then called monoicous. (Many sources, including those concerned with bryophytes, use the term 'monoecious' for this situation and 'dioecious' for the opposite. Here 'monoecious' and 'dioecious' are used only for sporophytes.)The liverwort Pellia epiphylla has the gametophyte as the dominant generation. It is monoicous: the small reddish sperm-producing antheridia are scattered along the midrib while the egg-producing archegonia grow nearer the tips of divisions of the plant.
Antheridia and archegonia occur on different gametophytes, which are then called dioicous.The moss Mnium hornum has the gametophyte as the dominant generation. It is dioicous: male plants produce only antheridia in terminal rosettes, female plants produce only archegonia in the form of stalked capsules. Seed plant gametophytes are also dioicous. However, the parent sporophyte may be monoecious, producing both male and female gametophytes or dioecious, producing gametophytes of one gender only. Seed plant gametophytes are extremely reduced in size; the archegonium consists only of a small number of cells, and the entire male gametophyte may be represented by only two cells.
Differentiation of the spores.
All spores the same size (homospory or isospory).Horsetails (species of Equisetum) have spores which are all of the same size.
Spores of two distinct sizes (heterospory or anisospory): larger megaspores and smaller microspores. When the two kinds of spore are produced in different kinds of sporangia, these are called megasporangia and microsporangia. A megaspore often (but not always) develops at the expense of the other three cells resulting from meiosis, which abort.
Megasporangia and microsporangia occur on the same sporophyte, which is then called monoecious.Most flowering plants fall into this category. Thus the flower of a lily contains six stamens (the microsporangia) which produce microspores which develop into pollen grains (the microgametophytes), and three fused carpels which produce integumented megasporangia (ovules) each of which produces a megaspore which develops inside the megasporangium to produce the megagametophyte. In other plants, such as hazel, some flowers have only stamens, others only carpels, but the same plant (i.e. sporophyte) has both kinds of flower and so is monoecious.
Megasporangia and microsporangia occur on different sporophytes, which are then called dioecious.An individual tree of the European holly (Ilex aquifolium) produces either 'male' flowers which have only functional stamens (microsporangia) producing microspores which develop into pollen grains (microgametophytes) or 'female' flowers which have only functional carpels producing integumented megasporangia (ovules) that contain a megaspore that develops into a multicellular megagametophyte.
There are some correlations between these variations, but they are just that, correlations, and not absolute. For example, in flowering plants, microspores ultimately produce microgametes (sperm) and megaspores ultimately produce megagametes (eggs). However, in ferns and their allies there are groups with undifferentiated spores but differentiated gametophytes. For example, the fern Ceratopteris thalictrioides has spores of only one kind, which vary continuously in size. Smaller spores tend to germinate into gametophytes which produce only sperm-producing antheridia.
A complex life cycle
The diagram shows the alternation of generations in a species which is heteromorphic, sporophytic, oogametic, dioicous, heterosporic and dioecious. A seed plant example might be a willow tree (most species of the genus Salix are dioecious). Starting in the centre of the diagram, the processes involved are:
An immobile egg, contained in the archegonium, fuses with a mobile sperm, released from an antheridium. The resulting zygote is either 'male' or 'female'.
A 'male' zygote develops by mitosis into a microsporophyte, which at maturity produces one or more microsporangia. Microspores develop within the microsporangium by meiosis.In a willow (like all seed plants) the zygote first develops into an embryo microsporophyte within the ovule (a megasporangium enclosed in one or more protective layers of tissue known as integument). At maturity, these structures become the seed. Later the seed is shed, germinates and grows into a mature tree. A 'male' willow tree (a microsporophyte) produces flowers with only stamens, the anthers of which are the microsporangia.
Microspores germinate producing microgametophytes; at maturity one or more antheridia are produced. Sperm develop within the antheridia.In a willow, microspores are not liberated from the anther (the microsporangium), but develop into pollen grains (microgametophytes) within it. The whole pollen grain is moved (e.g. by an insect or by the wind) to an ovule (megagametophyte), where a sperm is produced which moves down a pollen tube to reach the egg.
A 'female' zygote develops by mitosis into a megasporophyte, which at maturity produces one or more megasporangia. Megaspores develop within the megasporangium; typically one of the four spores produced by meiosis gains bulk at the expense of the remaining three, which disappear.'Female' willow trees (megasporophytes) produce flowers with only carpels (modified leaves that bear the megasporangia).
Megaspores germinate producing megagametophytes; at maturity one or more archegonia are produced. Eggs develop within the archegonia. The carpels of a willow produce ovules, megasporangia enclosed in integuments. Within each ovule, a megaspore develops by mitosis into a megagametophyte. An archegonium develops within the megagametophyte and produces an egg. The whole of the gametophytic 'generation' remains within the protection of the sporophyte except for pollen grains (which have been reduced to just three cells contained within the microspore wall).
Life cycles of different plant groups
The term "plants" is taken here to mean the Archaeplastida, i.e. the glaucophytes, red and green algae and land plants.
Alternation of generations occurs in almost all multicellular red and green algae, both freshwater forms (such as Cladophora) and seaweeds (such as Ulva). In most, the generations are homomorphic (isomorphic) and free-living. Some species of red algae have a complex triphasic alternation of generations, in which there is a gametophyte phase and two distinct sporophyte phases. For further information, see Red algae: Reproduction.
Land plants all have heteromorphic (anisomorphic) alternation of generations, in which the sporophyte and gametophyte are distinctly different. All bryophytes, i.e. liverworts, mosses and hornworts, have the gametophyte generation as the most conspicuous. As an illustration, consider a monoicous moss. Antheridia and archegonia develop on the mature plant (the gametophyte). In the presence of water, the biflagellate sperm from the antheridia swim to the archegonia and fertilisation occurs, leading to the production of a diploid sporophyte. The sporophyte grows up from the archegonium. Its body comprises a long stalk topped by a capsule within which spore-producing cells undergo meiosis to form haploid spores. Most mosses rely on the wind to disperse these spores, although Splachnum sphaericum is entomophilous, recruiting insects to disperse its spores. For further information, see Liverwort: Life cycle, Moss: Life cycle, Hornwort: Life cycle.
In ferns and their allies, including clubmosses and horsetails, the conspicuous plant observed in the field is the diploid sporophyte. The haploid spores develop in sori on the underside of the fronds and are dispersed by the wind (or in some cases, by floating on water). If conditions are right, a spore will germinate and grow into a rather inconspicuous plant body called a prothallus. The haploid prothallus does not resemble the sporophyte, and as such ferns and their allies have a heteromorphic alternation of generations. The prothallus is short-lived, but carries out sexual reproduction, producing the diploid zygote that then grows out of the prothallus as the sporophyte. For further information, see Fern: Life cycle.
In the spermatophytes, the seed plants, the sporophyte is the dominant multicellular phase; the gametophytes are strongly reduced in size and very different in morphology. The entire gametophyte generation, with the sole exception of pollen grains (microgametophytes), is contained within the sporophyte. The life cycle of a dioecious flowering plant (angiosperm), the willow, has been outlined in some detail in an earlier section (A complex life cycle). The life cycle of a gymnosperm is similar. However, flowering plants have in addition a phenomenon called 'double fertilization'. Two sperm nuclei from a pollen grain (the microgametophyte), rather than a single sperm, enter the archegonium of the megagametophyte; one fuses with the egg nucleus to form the zygote, the other fuses with two other nuclei of the gametophyte to form 'endosperm', which nourishes the developing embryo. For further information, see Double fertilization.
Evolution of the dominant diploid phase
It has been proposed that the basis for the emergence of the diploid phase of the life cycle (sporophyte) as the dominant phase (e.g. as in vascular plants) is that diploidy allows masking of the expression of deleterious mutations through genetic complementation. Thus if one of the parental genomes in the diploid cells contained mutations leading to defects in one or more gene products, these deficiencies could be compensated for by the other parental genome (which nevertheless may have its own defects in other genes). As the diploid phase was becoming predominant, the masking effect likely allowed genome size, and hence information content, to increase without the constraint of having to improve accuracy of DNA replication. The opportunity to increase information content at low cost was advantageous because it permitted new adaptations to be encoded. This view has been challenged, with evidence showing that selection is no more effective in the haploid than in the diploid phases of the lifecycle of mosses and angiosperms.
Similar processes in other organisms
Rhizaria
Some organisms currently classified in the clade Rhizaria and thus not plants in the sense used here, exhibit alternation of generations. Most Foraminifera undergo a heteromorphic alternation of generations between haploid gamont and diploid agamont forms. The single-celled haploid organism is typically much larger than the diploid organism.
Fungi
Fungal mycelia are typically haploid. When mycelia of different mating types meet, they produce two multinucleate ball-shaped cells, which join via a "mating bridge". Nuclei move from one mycelium into the other, forming a heterokaryon (meaning "different nuclei"). This process is called plasmogamy. Actual fusion to form diploid nuclei is called karyogamy, and may not occur until sporangia are formed. Karogamy produces a diploid zygote, which is a short-lived sporophyte that soon undergoes meiosis to form haploid spores. When the spores germinate, they develop into new mycelia.
Slime moulds
The life cycle of slime moulds is very similar to that of fungi. Haploid spores germinate to form swarm cells or myxamoebae. These fuse in a process referred to as plasmogamy and karyogamy to form a diploid zygote. The zygote develops into a plasmodium, and the mature plasmodium produces, depending on the species, one to many fruiting bodies containing haploid spores.
Animals
Alternation between a multicellular diploid and a multicellular haploid generation is never encountered in animals. In some animals, there is an alternation between parthenogenic and sexually reproductive phases (heterogamy). Both phases are diploid. This has sometimes been called "alternation of generations", but is quite different. In some other animals, such as hymenopterans, males are haploid and females diploid, but this is always the case rather than there being an alternation between distinct generations.
See also
: Evolutionary origin of the alternation of phases
Notes and references
Bibliography
Plant reproduction
Reproduction |
null | null | Aircraft pilot | eng_Latn | An aircraft pilot or aviator is a person who controls the flight of an aircraft by operating its directional flight controls. Some other aircrew members, such as navigators or flight engineers, are also considered aviators, because they are involved in operating the aircraft's navigation and engine systems. Other aircrew members, such as drone operators, flight attendants, mechanics and ground crew, are not classified as aviators.
In recognition of the pilots' qualifications and responsibilities, most militaries and many airlines worldwide award aviator badges to their pilots.
History
The first recorded use of the term aviator (aviateur in French) was in 1887, as a variation of "aviation", from the Latin avis (meaning bird), coined in 1863 by G. de la Landelle in Aviation Ou Navigation Aérienne ("Aviation or Air Navigation"). The term aviatrix (aviatrice in French), now archaic, was formerly used for a female aviator. These terms were used more in the early days of aviation, when airplanes were extremely rare, and connoted bravery and adventure. For example, a 1905 reference work described the Wright brothers' first airplane: "The weight, including the body of the aviator, is a little more than 700 pounds".
To ensure the safety of people in the air and on the ground, early aviation soon required that aircraft be under the operational control of a properly trained, certified pilot at all times, who is responsible for the safe and legal completion of the flight. The Aéro-Club de France delivered the first certificate to Louis Blériot in 1908—followed by Glenn Curtiss, Léon Delagrange, and Robert Esnault-Pelterie. The British Royal Aero Club followed in 1910 and the Aero Club of America in 1911 (Glenn Curtiss receiving the first).
Civilian
Civilian pilots fly aircraft of all types privately for pleasure, charity, or in pursuance of a business, or commercially for non-scheduled (charter) and scheduled passenger and cargo air carriers (airlines), corporate aviation, agriculture (crop dusting, etc.), forest fire control, law enforcement, etc. When flying for an airline, pilots are usually referred to as airline pilots, with the pilot in command often referred to as the captain.
Airline
There were 290,000 airline pilots in the world in 2017 and aircraft simulator manufacturer CAE Inc. forecasts a need for 255,000 new ones for a population of 440,000 by 2027, 150,000 for growth and 105,000 to offset retirement and attrition : 90,000 in Asia-Pacific (average pilot age in 2016: 45.8 years), 85,000 in Americas (48 years), 50,000 in Europe (43.7 years) and 30,000 in Middle East & Africa (45.7 years).
Boeing expects 790,000 new pilots in 20 years from 2018, 635,000 for commercial aviation, 96,000 for business aviation and 59,000 for helicopters: % in Asia Pacific (261,000), % in North America (206,000), % in Europe (146,000), % in the Middle East (64,000), % in Latin America (57,000), % in Africa (29,000) and % in Russia/ Central Asia (27,000).
By November 2017, due a shortage of qualified pilots, some pilots were leaving corporate aviation to return to airlines. In one example a Global 6000 pilot, making $250,000 a year for 10 to 15 flight hours a month, returned to American Airlines with full seniority.
A Gulfstream G650 or Global 6000 pilot might earn between $245,000 and $265,000, and recruiting one may require up to $300,000.
At the other end of the spectrum, constrained by the available pilots, some small carriers hire new pilots who need 300 hours to jump to airlines in a year. They may also recruit non-career pilots who have other jobs or airline retirees who want to continue to fly.
Automation
The number of airline pilots could decrease as automation replaces copilots and eventually pilots as well. In January 2017 Rhett Ross, CEO of Continental Motors said "my concern is that in the next two decades—if not sooner—automated and autonomous flight will have developed sufficiently to put downward pressure on both wages and the number and kind of flying jobs available. So if a kid asks the question now and he or she is 18, 20 years from now will be 2037 and our would-be careerist will be 38—not even mid-career. Who among us thinks aviation and especially for-hire flying will look like it does now?" Christian Dries, owner of Diamond Aircraft Austria said "Behind the curtain, aircraft manufacturers are working on a single-pilot cockpit where the airplane can be controlled from the ground and only in case of malfunction does the pilot of the plane interfere. Basically the flight will be autonomous and I expect this to happen in the next five to six years for freighters."
In August 2017 financial company UBS predicted pilotless airliners are technically feasible and could appear around 2025, offering around $35bn of savings, mainly in pilot costs: $26bn for airlines, $3bn for business jets and $2.1bn for civil helicopters; $3bn/year from lower pilot training and aviation insurance costs due to safer flights; $1bn from flight optimisation (1% of global airlines' $133bn jet fuel bill in 2016); not counting revenue opportunity from increased capacity utilization.
Regulations have to adapt with air cargo likely at the forefront, but pilotless flights could be limited by consumer behaviour: 54% of 8,000 people surveyed are defiant while 17% are supportive, with acceptation progressively forecast.
AVweb reporter Geoff Rapoport stated, "pilotless aircraft are an appealing prospect for airlines bracing for the need to hire several hundred thousand new pilots in the next decade. Wages and training costs have been rapidly rising at regional U.S. airlines over the last several years as the major airlines have hired pilots from the regionals at unprecedented rates to cover increased air travel demand from economic expansion and a wave of retirements".
Going to pilotless airliners could be done in one bold step or in gradual improvements like by reducing the cockpit crew for long haul missions or allowing single pilot cargo aircraft. The industry has not decided how to proceed yet. Present automated systems are not autonomous and must be monitored; their replacement could require artificial intelligence with machine learning while present certified software is deterministic.
As the Airbus A350 would only need minor modifications, Air Caraibes and French Bee parent Groupe Dubreuil see two-pilot crews in long-haul operations, without a third pilot for rotation, happening around 2024–2025.
Single-pilot freighters could start with regional flights.
The Air Line Pilots Association believe removing pilots would threaten aviation safety and opposes the April 2018 FAA Reauthorization Act's Section 744 establishing a research and development program to assist single-pilot cargo aircraft by remote and computer piloting.
For French aerospace research center Onera and avionics manufacturer Thales, artificial intelligence (AI) like consumer Neural Networks learning from large datasets cannot explain their operation and cannot be certified for safe air transport.
Progress towards ‘explainable’ AIs can be expected in the next decade, as the Onera expects “leads” for a certifiable AI system, along EASA standards evolution.
Africa and Asia
In some countries, such as Pakistan, Thailand and several African nations, there is a strong relationship between the military and the principal national airlines, and many airline pilots come from the military; however, that is no longer the case in the United States and Western Europe. While the flight decks of U.S. and European airliners do have ex-military pilots, many pilots are civilians. Military training and flying, while rigorous, is fundamentally different in many ways from civilian piloting.
Canada
Operating an aircraft in Canada is regulated by the Aeronautics Act of 1985 and the Canadian Aviation Regulations provide rules for Pilot licensing in Canada.
Retirement age is provided by each airline with some set to age 60, but changes to the Canadian Human Rights Act have restricted retirement age set by the airlines.
United States
In the United States in 2020, there were 691,691 active pilot certificates. This was down from a high of over 800,000 active pilots in 1980. Of the active pilot certificate holders, there were 160,860 Private, 103,879 Commercial, 164,193 Airline Transport, and 222,629 Student.
In 1930, the Air Commerce Act established pilot licensing requirements for American civil aviation.
Commercial airline pilots in the United States have a mandatory retirement age of 65, having increased from age 60 in 2007.
Military
Military pilots fly with the armed forces, primarily the air forces, of a government or nation-state. Their tasks involve combat and non-combat operations, including direct hostile engagements and support operations. Military pilots undergo specialized training, often with weapons. Examples of military pilots include fighter pilots, bomber pilots, transport pilots, test pilots and astronauts.
Military pilots are trained with a different syllabus than civilian pilots, which is delivered by military instructors. This is due to the different aircraft, flight goals, flight situations and chains of responsibility. Many military pilots do transfer to civilian-pilot qualification after they leave the military, and typically their military experience provides the basis for a civilian pilot's license.
Unmanned aerial vehicles
Unmanned aerial vehicles (UAVs, also known as "drones") operate without a pilot on-board and are classed into two categories: autonomous aircraft that operate without active human control during flight and remotely piloted UAVs which are operated remotely by one or more persons. The person controlling a remotely piloted UAV may be referred to as its pilot or operator. Depending on the sophistication and use of the UAV, pilots/operators of UAVs may require certification or training, but are generally not subject to the licensing/certification requirements of pilots of manned aircraft.
Most jurisdictions have restrictions on the use of UAVs which have greatly limited their use in controlled airspace; UAVs have mostly been limited to military and hobbyist use. In the United States, use of UAVs is very limited in controlled airspace (generally, above 400 ft/122m and away from airports) and the FAA prohibits nearly all commercial use. Once regulations are made to allow expanded use of UAVs in controlled airspace, there is expected to be a large surge of UAVs in use and, consequently, high demand for pilots/operators of these aircraft.
Space
The general concept of an airplane pilot can be applied to human spaceflight, as well. The spacecraft pilot is the astronaut who directly controls the operation of a spacecraft. This term derives directly from the usage of the word "pilot" in aviation, where it is synonymous with "aviator".
Pilot certifications
Pilots are required to go through many hours of flight training and theoretical study, that differ depending on the country. The first step is acquiring the Private Pilot License (PPL), or Private Pilot Certificate. In the United States of America, this includes a minimum of 35 to 40 hours of flight training, the majority of which with a Certified Flight Instructor.
In the United States, an LSA (Light Sport Aircraft) license can be obtained in at least 20 hours of flight time.
The next step in a pilot's progression is either Instrument Rating (IR), or Multi-Engine Rating (MEP) endorsements.
If a professional career or professional-level skills are desired, a Commercial Pilot License (CPL) would also be required. To captain an airliner, one must obtain an Airline Transport Pilot License (ATPL). After 1 August 2013, even when being a First Officer (FO), an ATPL is required. Some countries/carriers require/use a Multi Crew Coordination (MCC).
See also
Airline pilot uniforms
Air safety
IMSAFE (mnemonic for pilot's fitness to fly)
List of aerospace engineers
List of aviators
List of Russian aviators
Pilot fatigue
Pilot logbook
References
External links
U.S. Women Pilots Statistics 1960–2010
Military aviation occupations
Occupations in aviation |
null | null | Colossus of Rhodes | eng_Latn | The Colossus of Rhodes ( ) was a statue of the Greek sun-god Helios, erected in the city of Rhodes, on the Greek island of the same name, by Chares of Lindos in 280 BC. One of the Seven Wonders of the Ancient World, it was constructed to celebrate the successful defence of Rhodes city against an attack by Demetrius Poliorcetes, who had besieged it for a year with a large army and navy.
According to most contemporary descriptions, the Colossus stood approximately 70 cubits, or high – approximately the height of the modern Statue of Liberty from feet to crown – making it the tallest statue in the ancient world. It collapsed during the earthquake of 226 BC, although parts of it were preserved. In accordance with a certain oracle, the Rhodians did not build it again. John Malalas wrote that Hadrian in his reign re-erected the Colossus, but he was mistaken. According to the Suda, the Rhodians were called Colossaeans (Κολοσσαεῖς), because they erected the statue on the island.
In 653, an Arab force under Muslim general Muawiyah I conquered Rhodes, and according to the Chronicle of Theophanes the Confessor, the statue was completely destroyed and the remains sold.
Since 2008, a series of as-yet-unrealized proposals to build a new Colossus at Rhodes Harbour have been announced, although the actual location of the original monument remains in dispute.
Siege of Rhodes
In the early fourth century BC, Rhodes, allied with Ptolemy I of Egypt, prevented a mass invasion staged by their common enemy, Antigonus I Monophthalmus.
In 304 BC a relief force of ships sent by Ptolemy arrived, and Demetrius (son of Antigonus) and his army abandoned the siege, leaving behind most of their siege equipment. To celebrate their victory, the Rhodians sold the equipment left behind for 300 talents and decided to use the money to build a colossal statue of their patron god, Helios. Construction was left to the direction of Chares, a native of Lindos in Rhodes, who had been involved with large-scale statues before. His teacher, the sculptor Lysippos, had constructed a bronze statue of Zeus at Tarentum.
Construction
Construction began in 292 BC. Ancient accounts, which differ to some degree, describe the structure as being built with iron tie bars to which brass plates were fixed to form the skin. The interior of the structure, which stood on a white marble pedestal near the Rhodes harbor entrance, was then filled with stone blocks as construction progressed. Other sources place the Colossus on a breakwater in the harbor. According to most contemporary descriptions, the statue itself was about 70 cubits, or tall. Much of the iron and bronze was reforged from the various weapons Demetrius's army left behind, and the abandoned second siege tower may have been used for scaffolding around the lower levels during construction.
Philo of Byzantium wrote in De septem mundi miraculis that Chares created the sculpture in situ by casting it in horizontal courses and then placing "...a huge mound of earth round each section as soon as it was completed, thus burying the finished work under the accumulated earth, and carrying out the casting of the next part on the level."
Modern engineers have put forward a plausible hypothesis for the statue's construction, based on the technology of the time (which was not based on the modern principles of earthquake engineering), and the accounts of Philo and Pliny, who saw and described the ruins.
The base pedestal was said to be at least in diameter, and either circular or octagonal. The feet were carved in stone and covered with thin bronze plates riveted together. Eight forged iron bars set in a radiating horizontal position formed the ankles and turned up to follow the lines of the legs while becoming progressively smaller. Individually cast curved bronze plates square with turned-in edges were joined together by rivets through holes formed during casting to form a series of rings. The lower plates were in thickness to the knee and thick from knee to abdomen, while the upper plates were thick except where additional strength was required at joints such as the shoulder, neck, etc.
Archaeologist Ursula Vedder has proposed that the sculpture was cast in large sections following traditional Greek methods and that Philo's account is "not compatible with the situation proved by archaeology in ancient Greece."
The Standing Colossus (280–226 BC)
After twelve years, in 280 BC, the statue was completed. Preserved in Greek anthologies of poetry is what is believed to be the genuine dedication text for the Colossus.
Collapse (226 BC)
The statue stood for 54 years until a 226 BC earthquake caused significant damage to large portions of Rhodes, including the harbour and commercial buildings, which were destroyed. The statue snapped at the knees and fell over onto land. Ptolemy III offered to pay for the reconstruction of the statue, but the oracle of Delphi made the Rhodians fear that they had offended Helios, and they declined to rebuild it.
Fallen state (226 BC to 653 AD)
The remains lay on the ground for over 800 years, and even broken, they were so impressive that many travelled to see them.
The remains were described briefly by Strabo (64 or 63 BC – c. AD 24), in his work Geography (Book XIV, Chapter 2.5). Strabo was a Greek geographer, philosopher, and historian who lived in Asia Minor during the transitional period of the Roman Republic into the Roman Empire.
Strabo is best known for his work Geographica ("Geography"), which presented a descriptive history of people and places from different regions of the world known during his lifetime. Strabo states that:
Pliny the Elder (AD 23/24 – 79) was a Roman author, a naturalist and natural philosopher, a naval and army commander of the early Roman Empire, and a friend of emperor Vespasian. Pliny wrote the encyclopedic Naturalis Historia (Natural History), which became an editorial model for encyclopedias. The Naturalis Historia is one of the largest single works to have survived from the Roman Empire to the modern day and purports to cover the entire field of ancient knowledge. Pliny remarked:
{{quote|text= But that which is by far the most worthy of our admiration, is the colossal statue of the Sun, which stood formerly at Rhodes, and was the work of Chares the Lindian, a pupil of the above-named Lysippus; no less than seventy cubits in height. This statue fifty-six years after it was erected, was thrown down by an earthquake; but even as it lies, it excites our wonder and admiration. Few men can clasp the thumb in their arms, and its fingers are larger than most statues. Where the limbs are broken asunder, vast caverns are seen yawning in the interior. Within it, too, are to be seen large masses of rock, by the weight of which the artist steadied it while erecting it.<ref>Natural History, book 34, Natural History of Metals xviii, 41.</ref>}}
Destruction of the remains (653)
In 653, an Arab force under Muslim general Muawiyah I captured Rhodes, and according to the Chronicle of Theophanes the Confessor, the statue was melted down and sold to a Jewish merchant of Edessa who loaded the bronze onto 900 camels. The Arab destruction and the purported sale to a Jew possibly originated as a powerful metaphor for Nebuchadnezzar's dream of the destruction of a great statue.
The same story is recorded by Bar Hebraeus, writing in Syriac in the 13th century in Edessa (after the Arab pillage of Rhodes): "And a great number of men hauled on strong ropes which were tied round the brass Colossus which was in the city and pulled it down. And they weighed from it three thousand loads of Corinthian brass, and they sold it to a certain Jew from Emesa" (the Syrian city of Homs). Theophanes is the sole source of this account, and all other sources can be traced to him.
Posture
The harbour-straddling Colossus was a figment of medieval imaginations based on the dedication text's mention of "over land and sea" twice and the writings of an Italian visitor who in 1395 noted that local tradition held that the right foot had stood where the church of St John of the Colossus was then located. Many later illustrations show the statue with one foot on either side of the harbour mouth with ships passing under it. References to this conception are also found in literary works. Shakespeare's Cassius in Julius Caesar (I, ii, 136–38) says of Caesar:
Shakespeare alludes to the Colossus also in Troilus and Cressida (V.5) and in Henry IV, Part 1 (V.1).
"The New Colossus" (1883), a sonnet by Emma Lazarus written on a cast bronze plaque and mounted inside the pedestal of the Statue of Liberty in 1903, contrasts the latter with:
While these fanciful images feed the misconception, the mechanics of the situation reveal that the Colossus could not have straddled the harbour as described in Lemprière's Classical Dictionary. If the completed statue had straddled the harbour, the entire mouth of the harbour would have been effectively closed during the entirety of the construction, and the ancient Rhodians did not have the means to dredge and re-open the harbour after construction. Also, the fallen statue would have blocked the harbour, and since the ancient Rhodians did not have the ability to remove the fallen statue from the harbour, it would not have remained visible on land for the next 800 years, as discussed above. Even neglecting these objections, the statue was made of bronze, and engineering analyses indicate that it could not have been built with its legs apart without collapsing under its own weight. Many researchers have considered alternative positions for the statue which would have made it more feasible for actual construction by the ancients. There is also no evidence that the statue held a torch aloft; the records simply say that after completion, the Rhodians kindled the "torch of freedom". A relief in a nearby temple shows Helios standing with one hand shielding his eyes, similar to the way a person shields their eyes when looking toward the sun, and it is quite possible that the colossus was constructed in the same pose.
While scholars do not know what the statue looked like, they do have a good idea of what the head and face looked like, as it was of a standard rendering at the time. The head would have had curly hair with evenly spaced spikes of bronze or silver flame radiating, similar to the images found on contemporary Rhodian coins.
Possible locations
While scholars generally agree that anecdotal depictions of the Colossus straddling the harbour's entry point have no historic or scientific basis, the monument's actual location remains a matter of debate. As mentioned above the statue is thought locally to have stood where two pillars now stand at the Mandraki port entrance.
The floor of the Fortress of St Nicholas, near the harbour entrance, contains a circle of sandstone blocks of unknown origin or purpose. Curved blocks of marble that were incorporated into the Fortress structure, but are considered too intricately cut to have been quarried for that purpose, have been posited as the remnants of a marble base for the Colossus, which would have stood on the sandstone block foundation.
Archaeologist Ursula Vedder postulates that the Colossus was not located in the harbour area at all, but rather was part of the Acropolis of Rhodes, which stood on a hill that overlooks the port area. The ruins of a large temple, traditionally thought to have been dedicated to Apollo, are situated at the highest point of the hill. Vedder believes that the structure would actually have been a Helios sanctuary, and a portion of its enormous stone foundation could have served as the supporting platform for the Colossus.
Modern Colossus projects
In 2008, The Guardian reported that a modern Colossus was to be built at the harbour entrance by the German artist Gert Hof leading a Cologne-based team. It was to be a giant light sculpture made partially out of melted-down weapons from around the world. It would cost up to €200 million.
In December 2015, a group of European architects announced plans to build a modern Colossus bestriding two piers at the harbour entrance, despite a preponderance of evidence and scholarly opinion that the original monument could not have stood there. The new statue, tall (five times the height of the original), would cost an estimated US$283 million, funded by private donations and crowdsourcing. The statue would include a cultural centre, a library, an exhibition hall, and a lighthouse, all powered by solar panels. , no such plans have been carried out and the website for the project is offline.
In popular cultureThe Colossus of Rhodes, a 1961 film, depicts Rhodes in the Hellenistic period. It was the directorial debut of Sergio Leone.
On the cartoon show DuckTales, the episode "Home Sweet Homer" (production code 130, originally airing 23 October 1987), the characters talk about the lost city of Ithaquack, which used to have The Colossus of Duckapopolis straddling its entrance (much as the Colossus of Rhodes is thought to have been poised) to guide ships to the otherwise unseen city.
The Colossus is featured in the 2007 video game God of War II. In an attempt to kill the game's protagonist, Kratos, Zeus uses magic to animate the unfinished Colossus. Kratos defeats the statue by attacking it from within, causing its head to explode, but gets caught and injured underneath its falling hand while boasting about his victory.
In George R.R. Martin's fantasy series A Song of Ice and Fire, the city of Braavos is located in a lagoon guarded by a giant fortress in the shape of a man. This 'titan of Braavos', in reference to the Colossus, straddles the bay leading to the city.
The song Colossus by Swedish death metal band Avatar, on the 2020 album Hunter Gatherer, is confirmed by lead singer Johannes Eckerström to have been inspired by the Colossus of Rhodes, as well as named after it.
The Colossus appears in Omega, the sixth main book in Jeremy Robinson’s Chess Team series. In 226 B.C., the statue mysteriously comes to life and ravages the island of Rhodes before collapsing to the ground, where it remained for centuries to come. In 2013, the Colossus’s remains have somehow been relocated to the Gulf of Tunisia, submerged underwater near the Omega Manifold facility. It is here that rogue geneticist Richard Ridley uses the mother tongue to bring the Colossus to life, intending to use it as part of his plan to achieve world domination.
See also
Twelve Metal Colossi
The Colossus of Rhodes (Dalí)The New ColossusThe Rhodes ColossusList of tallest statues
References
Notes
References
Sources
Gabriel, Albert. Bulletin de Correspondance Hellenique 56 (1932), pp. 331–59.
Haynes, D.E.L. "Philo of Byzantium and the Colossus of Rhodes" The Journal of Hellenic Studies 77.2 (1957), pp. 311–312. A response to Maryon.
Maryon, Herbert, "The Colossus of Rhodes" in The Journal of Hellenic Studies 76 (1956), pp. 68–86. A sculptor's speculations on the Colossus of Rhodes.
Further reading
Jones, Kenneth R. 2014. "Alcaeus of Messene, Philip V and the Colossus of Rhodes: A Re-Examination of Anth. Pal. 6.171." The Classical Quarterly 64, no. 1: 136–51. doi:10.1017/S0009838813000591.
Romer, John., and Elizabeth Romer. 1995. The Seven Wonders of the World: A History of the Modern Imagination. 1st American ed. New York: Henry Holt.
Woods, David. 2016. "On the Alleged Arab Destruction of the Colossus of Rhodes c. 653." Byzantion: Revue Internationale Des Etudes Byzantines'' 86: 441–51.
Ancient Rhodes
Demolished buildings and structures in Greece
Seven Wonders of the Ancient World
Ancient Greece in art and culture
Hellenistic architecture
Hellenistic and Roman bronzes
Lost sculptures
Culture of Rhodes
Helios
Ancient Greek metalwork
Buildings and structures in Rhodes (city)
3rd-century BC religious buildings and structures
Buildings and structures completed in the 3rd century BC
Buildings and structures demolished in the 3rd century BC
3rd-century BC establishments
3rd-century BC disestablishments
Ancient Greek and Roman colossal statues |
null | null | Under Siege 2: Dark Territory | eng_Latn | Under Siege 2: Dark Territory is a 1995 American action thriller film directed by Geoff Murphy, starring Steven Seagal as the ex-Navy SEAL, Casey Ryback. Set on board a train traveling through the Rocky Mountains from Denver to Los Angeles, it is the sequel to the 1992 film Under Siege also starring Seagal. The title refers to the railroading term that the subject train was travelling through dark territory, a section of railroad track that has no train signals and in which communications between train dispatchers and the railroad engineers were impossible.
The film was produced by Seagal along with Arnon Milchan and Steve Perry. The film's cast also included Eric Bogosian, Everett McGill, Morris Chestnut, Peter Greene, Kurtwood Smith and Katherine Heigl. In addition to Seagal, Nick Mancuso, Andy Romano, and Dale Dye also reprised their roles from the first film.
Plot
Following his retirement from the United States Navy, Casey Ryback settles in Denver, Colorado, where he starts and runs a restaurant business. Sometime later, his estranged brother, James Ryback, perishes in a plane crash. Upon receiving news of the incident, Casey meets James's daughter, Sarah, whom he will accompany to Los Angeles to attend his funeral. The two board the Grand Continental, a train traveling through the Rocky Mountains from Denver to Los Angeles. On board, they befriend a porter named Bobby Zachs and the train's chefs.
As the train reaches the Rocky Mountains, a group of terrorists flag it down and murder the engineer and brakeman. The group, led by former U.S. government employee and computer genius Travis Dane and his second-in-command Marcus Penn, cuts the train phone lines and takes passengers and staff hostage, herding them into the last two cars. Casey kills one terrorist, then slips away. Dane worked on Grazer One, a top secret military satellite particle weapon for underground targets. The military fired Dane, who later faked his suicide.
Dane threatens two former Department of Defense colleagues with burning needles in their eyes unless they reveal codes to take over Grazer. Despite their disclosures, after the codes are confirmed to work, they are thrown from the train over a deep valley.
Middle Eastern terrorists have offered Dane $1 billion to destroy the Eastern seaboard by using Grazer to target a nuclear reactor located under the Pentagon. Dane demonstrates Grazer to investors by destroying a Chinese waste-processing plant. After one investor offers an additional $100 million, despite the weapon's primary intention for subterranean targets, Dane destroys an airliner carrying the investor's ex-wife.
The U.S. government has difficulty locating Dane or Grazer. When officials destroy what they think is Grazer, Dane explains the NSA's best intelligence satellite was destroyed. As long as the train keeps moving, his location cannot be determined.
With Zachs' help, Casey takes matters into his own hands. He faxes a message to the owner of the Mile High Cafe, who relays the word to Admiral Bates. Bates quickly understands that Dane and the terrorists are on the train, and reluctantly approves a mission by two F-117 stealth bombers to destroy the train.
Zachs discovers that they are on the wrong tracks and are on a collision course with a Southern Pacific bulk freight train carrying gasoline tank cars. Casey kills the mercenaries one by one and releases the hostages, but Dane uses his computer skills to find the Stealth bombers and re-target Grazer to knock them out before they complete their mission. Aware of Casey's past, Penn uses Sarah as a bait to lure him to a fight, but Casey gains the upper hand and fatally breaks his neck.
Casey finds Dane about to depart in a chopper hovering over the train. When Dane informs Casey that there is no way to stop the satellite from destroying Washington, Casey shoots him. The bullet destroys his computer and injures Dane, who falls out of a window of the train. Control of the satellite is restored at the Pentagon and it is destroyed by remote control one second before it would have fired on the Pentagon, just before the two trains collide.
The crash happens on a trestle, resulting in an explosion that destroys the bridge and kills Scotty, the mercenary driving the train. Casey races through the exploding train, and grabs a rope ladder hanging from the chopper. Dane, who survived the gunshot and the crash, also catches onto the ladder. He attempts to climb onto the helicopter, screaming that he and Casey should join forces. Casey shuts the door, severing Dane's fingers and the latter falls to his death.
Casey informs the Pentagon that the passengers are safe, having previously detached the last two cars from the rest of the train. Later on, Sarah and Casey pay their final respects at James' gravestone.
Cast
Steven Seagal as Lieutenant Casey Ryback, a former Navy SEAL who now heads and manages a restaurant in Denver
Eric Bogosian as Travis Dane, a crazed computer genius, cyberterrorist leader who designed the Grazer One satellite weapon for the US government before being dismissed for his mental instability
Everett McGill as Marcus Penn, the mercenary commander & Dane's second-in-command who leads a team of mercenaries to hijack the Grand Continental train to set up the satellite equipment
Katherine Heigl as Sarah Ryback, Casey's estranged niece whom he accompanies on the journey to her father's funeral
Morris Chestnut as Bobby Zachs, an eager porter of the train who reluctantly helps Ryback with the hijacked train
Nick Mancuso as Tom Breaker, the shady CIA director who assists ATAC on Grazer One
Brenda Bakke as Captain Linda Gilder, a member of ATAC and one of Dane's former colleagues
Peter Greene as Mercenary #1, Penn's lieutenant who was once instructed by Ryback at Fort Bragg
Patrick Kilpatrick as Mercenary #2, One of Penn's mercenaries
Scott Sowers as Mercenary #3, One of Penn's mercenaries
Afifi Alaouie as Fatima, the sole female of Penn's mercenaries
Andy Romano as Admiral Bates, Chairman of the Joint Chiefs of Staff
Dale Dye as Captain Nick Garza, Admiral Bates' right-hand man
Kurtwood Smith as Major General Stanley Cooper, an Air Force general who commands ATAC and Dane's former boss
David Gianopoulos as Captain David Trilling, a member of ATAC and one of Dane's former colleagues
Sandra Taylor as Kelly, a barmaid on board the train
Jonathan Banks as Scotty, the mercenary driver under the orders of Penn to drive the locomotive of the Grand Continental train
Royce D. Applegate as Ryback's cook who runs the restaurant when Ryback is absent
Dale Payne as Train Conductor
Rail equipment, railroad companies and rail lines featured
Rail lines by company and equipment
Production
Development
The film was an early credit for Matt Reeves who wrote the script with a friend of his in college. Reeves later said, "There was a big action spec market, a lot of movies that were selling, and so we wrote this movie with my thought being, ‘And then I will be able to finance my student film and that way I can become a director’."
The script was called Dark Territory at one stage and End of the Line at another. When they finished writing it "the spec market crashed and it didn't sell." But the script was optioned by Warner Brothers, who decided to turn the film into a sequel to Under Siege.
Reeves said the film was originally "meant to be very much like a Die Hard movie, which I guess Under Siege really was too, except the difference was that in the Under Siege movies that tension is how soon before Seagal will rip out someone's larynx. And what I love about Die Hard was this idea of the underdog, that here's this guy, especially in that first movie, who's a cop from New York who doesn't even have shoes. And somehow he has got to save this building, save the day. That was what that movie was supposed to be, but it didn't end up being that."
Filming
According to Morris Chestnut Seagal rewrote many of the scenes he was in. "The only time they really stuck to the script or had ad libs was the stuff when he really wasn't there. It was a lot of stuff, because at that time I think he was flying a helicopter, he was doing something... He would come to set, "Okay, you're gonna say this. I'm gonna say this and this is gonna happen and then you do that." That's how we did a lot of that movie."
Part of the film was shot in Chatsworth's Stony Point Park. The production painted some of the boulders, which upset rock climbers who claimed it made them unsafe.
The locomotive seen in the film had previously appeared in the 1985 film Runaway Train.
Director Geoff Murphy called making the film "a very dreary process and very highly contentious at the time – lots of arguments and stuff. There was a point during the editing where I observed this incredibly high energy beast emerging, and I didn't know where it had come from, because there wasn't any of that energy on the set. It seemed to grow out of the editing process."
Reception
Box office
Under Siege 2 opened at #2 at the box office under Apollo 13 in 2,150 theaters and made $12,624,402 for the weekend.
Critical response
On Rotten Tomatoes, the review aggregation website, the film has an approval rating of 34% based on 32 reviews and an average rating of 4.73/10. The site's critical consensus states, "Utterly forgettable and completely unnecessary, Under Siege 2 represents a steep comedown from its predecessor – and an unfortunate return to form for its star." On Metacritic, the film has a score of 52 out of 100 based on reviews from 21 critics, indicating "mixed or average reviews". Audiences polled by CinemaScore gave the film an average grade of "A-" on an A+ to F scale.
Roger Ebert of the Chicago Sun-Times gave the film a three-star rating in his review, while Peter Rainer of the Los Angeles Times wrote that "the action upstaged the actors." Leonard Klady of Variety magazine notes Seagal's confidence but says he is betrayed by his limited performance. Klady praises the villains for their performances, but says "they and others are saddled with pedestrian dialogue and motivation."
Legacy
Seagal was later criticized for his behavior during the film.
Jenny McCarthy unsuccessfully auditioned for a role in the film. She said Seagal auditioned her and asked her to take off her clothes.
Katherine Heigl, who was sixteen at the time the film was made, said that on the last day of filming Seagal told her, "'You know Katie, I got girlfriends your age.' And I said, 'Isn't that illegal?' And he said, 'They don't seem to mind'."
Sequel
On October 3, 2016, Seagal announced on Twitter that the script for Under Siege 3 was in development.
Notes
References
External links
1995 films
1995 action thriller films
1990s English-language films
American films
American action thriller films
American sequel films
Films about hijackings
Films about terrorism
Films directed by Geoff Murphy
Films scored by Basil Poledouris
Films set in Colorado
Films set on trains
Films with screenplays by Matt Reeves
Regency Enterprises films
Siege films
Warner Bros. films |
null | null | Bullet | eng_Latn | A bullet is a kinetic projectile, a component of firearm ammunition that is shot from a gun barrel. The term is from Middle French, originating as the diminutive of the word boulle (boullet), which means "small ball". Bullets are made of a variety of materials, such as copper, lead, steel, polymer, rubber and even wax. Bullets are made in various shapes and constructions (depending on the intended applications), including specialized functions such as hunting, target shooting, training and combat. Bullets are often tapered, making them more aerodynamic. Bullet sizes are expressed by their weights and diameters (referred to as "calibers") in both imperial and metric measurement systems. For example: 55 grain .223 caliber bullets are of the same weight and caliber as 3.56 gram 5.56mm caliber bullets. Bullets do not normally contain explosives (see Incendiary ammunition and Exploding bullet), but strike or damage the intended target by transferring kinetic energy upon impact and penetration (see terminal ballistics).
Bullets are available singly (as in muzzle-loading and cap and ball firearms), but are more often packaged with propellant as cartridges ("rounds" of ammunition). Bullets are components of paper cartridges, or (much more commonly) in the form of metallic cartridges. Although the word bullet is often used in colloquial language to refer to a cartridge round, a bullet is not a cartridge but rather a component of one. A cartridge is a combination package of the bullet (i.e., the projectile), the case (which holds everything together), the propellant (which provides the majority of the energy to launch the projectile) and the primer (which ignites the propellant). This use of the term bullet (when intending to describe a cartridge) often leads to confusion when a cartridge, and all its components, are specifically referred to. The cartridges, in turn, may be held in a magazine or a belt (for rapid-fire weapons).
The bullets used in many cartridges are fired at muzzle velocities faster than the speed of sound—about in dry air at —and thus can travel a substantial distance to a target before a nearby observer hears the sound of the shot. The sound of gunfire (i.e. the "muzzle report") is often accompanied with a loud bullwhip-like crack as the supersonic bullet pierces through the air creating a sonic boom. Bullet speeds at various stages of flight depend on intrinsic factors such as sectional density, aerodynamic profile and ballistic coefficient, and extrinsic factors such as barometric pressure, humidity, air temperature and wind speed. Subsonic cartridges fire bullets slower than the speed of sound, so there are no sonic booms. This means that a subsonic cartridge, such as .45 ACP, can be substantially quieter than a supersonic cartridge, such as the .223 Remington, even without the use of a suppressor.
Bullets shot by firearms can be used for target practice or to injure or kill animals, or people. Death can be by blood loss or damage to vital organs, or even asphyxiation if blood enters the lungs. Bullets are not the only projectiles shot from firearm-like equipment: BBs are shot from BB guns, airsoft pellets are shot by airsoft guns, paintballs are shot by paintball markers, and small rocks can be hurtled from slingshots. There are also flare guns, potato guns (and spud guns), rubber bullets, tasers, bean bag rounds, grenade launchers, flash bangs, tear gas, RPGs, and missile launchers.
History
While the first recorded use of gunpowder in Europe was in 1247, it was invented in China during the 9th century. The cannon appeared in 1327, and the hand cannon in 1364. Early projectiles were made of stone. Eventually it was discovered that stone would not penetrate stone fortifications, which led to the use of denser materials as projectiles. Hand cannon projectiles developed in a similar manner. The first recorded instance of a metal ball from a hand cannon penetrating armor was in 1425. Shot retrieved from the wreck of the Mary Rose (sunk in 1545, raised in 1982) are of different sizes, and some are stone while others are cast iron.
The development of the hand culverin and matchlock arquebus brought about the use of cast lead balls as projectiles. The original round musket ball was smaller than the bore of the barrel. At first it was loaded into the barrel just resting upon the powder. Later, some sort of material was used as a wadding between the ball and the powder as well as over the ball to keep it in place, it held the bullet firmly in the barrel and against the powder. (Bullets not firmly on the powder risked exploding the barrel, with the condition known as a "short start".)
The loading of muskets was, therefore, easy with the old smooth-bore Brown Bess and similar military muskets. The original muzzle-loading rifle, however, was loaded with a piece of leather or cloth wrapped around the ball, to allow the ball to engage the grooves in the barrel. Loading was a bit more difficult, particularly when the bore of the barrel was fouled from previous firings. For this reason, and because rifles were not often fitted for bayonets, early rifles were rarely used for military purposes, compared to muskets.
The first half of the nineteenth century saw a distinct change in the shape and function of the bullet. In 1826, Henri-Gustave Delvigne, a French infantry officer, invented a breech with abrupt shoulders on which a spherical bullet was rammed down until it caught the rifling grooves. Delvigne's method, however, deformed the bullet and was inaccurate.
Square bullets have origins that almost pre-date civilization and were used by slingers in slings. They were typically made out of copper or lead. The most notable use of square bullet designs was done by James Puckle and Kyle Tunis who patented them, where they were briefly used in one version of the Puckle gun. The early use of these in the black-powder era was soon discontinued due to irregular and unpredictable flight patterns.
Pointed bullets
Delvigne continued to develop bullet design and by 1830 had started to develop cylindro-conical bullets. His bullet designs were improved by Francois Tamisier with the addition of "ball grooves" which are known as "cannelures", which moved the resistance of air behind the center of gravity of the bullet.
Tamisier also developed progressive rifling. The rifle grooves were deeper toward the breech, becoming shallower as they progressed toward the muzzle. This causes the bullet to be progressively molded into the grooves which increases range and accuracy.
The Thouvenin rifle barrel has a forcing plug in the breech of the barrel to mold the bullet into the rifling with the use of a special ramrod. While successful in increasing accuracy it was extremely hard to clean. These improvements were the basis for the development of the Minié ball.
Among the first pointed or "conical" bullets were those designed by Captain John Norton of the British Army in 1832. Norton's bullet had a hollow base made of lotus pith that, on firing, expanded under pressure to engage with a barrel's rifling. The British Board of Ordnance rejected it because spherical bullets had been in use for the previous 300 years.
Renowned English gunsmith William Greener invented the Greener bullet in 1836. Greener fitted the hollow base of an oval bullet with a wooden plug that more reliably forced the base of the bullet to expand and catch the rifling. Tests proved that Greener's bullet was extremely effective, but the military rejected it because, being two parts, they judged it as too complicated to produce.
The soft lead Minié ball was first introduced in 1847 by Claude-Étienne Minié, a captain in the French Army. It was another improvement of the work done by Delvigne. As designed by Minié, the bullet was conical in shape with a hollow cavity in the rear, which was fitted with a small iron cap instead of a wooden plug. When fired, the iron cap forced itself into the hollow cavity at the rear of the bullet, thus expanding the sides of the bullet to grip and engage the rifling. In 1851, the British adopted the Minié ball for their 702-inch Pattern 1851 Minié rifle. In 1855, James Burton, a machinist at the US Armory at Harper's Ferry, West Virginia, improved the Minié ball further by eliminating the metal cup in the bottom of the bullet. The Minié ball first saw widespread use in the Crimean War (1853-1856). Roughly 90% of the battlefield casualties in the American Civil War (1861-1865) were caused by Minié balls fired from rifled muskets.
A similar bullet called the Nessler ball was also developed for smoothbore muskets.
Between 1854 and 1857, Sir Joseph Whitworth conducted a long series of rifle experiments, and proved, among other points, the advantages of a smaller bore and, in particular, of an elongated bullet. The Whitworth bullet was made to fit the grooves of the rifle mechanically. The Whitworth rifle was never adopted by the government, although it was used extensively for match purposes and target practice between 1857 and 1866, when it was gradually superseded by Metford's.
In 1861 W.B. Chace approached President Abraham Lincoln with an improved ball design for muskets. In firing over the Potomac river where the Chace ball and the round ball were alternated Lincoln observed that the Chace design carried a third or more farther fired at the same elevation. Although Lincoln recommended testing, it never took place.
About 1862 and later, W. E. Metford carried out an exhaustive series of experiments on bullets and rifling, and invented the important system of light rifling with increasing spiral, and a hardened bullet. The combined result was that, in December 1888, the Lee–Metford small-bore (.303", 7.70 mm) rifle, Mark I, was finally adopted for the British army. The Lee–Metford was the predecessor of the Lee–Enfield.
Modern bullets
The next important change in the history of the rifle bullet occurred in 1882, when Lt. Colonel Eduard Rubin, director of the Swiss Army Laboratory at Thun, invented the copper-jacketed bullet — an elongated bullet with a lead core in a copper jacket. It was also small bore (7.5mm and 8mm) and it is the precursor of the 8mm Lebel bullet adopted for the smokeless powder ammunition of the Lebel Model 1886 rifle.
The surface of lead bullets fired at high velocity may melt due to hot gases behind and friction with the bore. Because copper has a higher melting point, and greater specific heat capacity and hardness, copper-jacketed bullets allow greater muzzle velocities.
European advances in aerodynamics led to the pointed spitzer bullet. By the beginning of the twentieth century, most world armies had begun the transition to spitzer bullets. These bullets flew for greater distances more accurately and carried more energy with them. Spitzer bullets combined with machine guns greatly increased the lethality of the battlefield.
The latest advancement in bullet shape was the boat tail, a streamlined base for spitzer bullets. The vacuum created as air moving at high speed passes over the end of a bullet slows the projectile. The streamlined boat tail design reduces this form drag by allowing the air to flow along the surface of the tapering end. The resulting aerodynamic advantage is currently seen as the optimum shape for rifle technology. The first combination spitzer and boat-tail bullet, named Balle "D" from its inventor (a lieutenant-colonel Desaleux), was introduced as standard military ammunition in 1901, for the French Lebel Model 1886 rifle.
A ballistic tip bullet is a hollow-point rifle bullet that has a plastic tip on the end of the bullet itself. This improves external ballistics by streamlining the bullet, allowing it to cut through the air more easily, and improves terminal ballistics by allowing the bullet to act as a JHP on impact.
As a side effect, it also feeds better in weapons that have trouble feeding rounds that are not FMJ rounds.
Design
Bullet designs have to solve two primary problems. In the barrel, they must first form a seal with the gun's bore. If a strong seal is not achieved, gas from the propellant charge leaks past the bullet, thus reducing efficiency and possibly accuracy. The bullet must also engage the rifling without damaging or excessively fouling the gun's bore, and without distorting the bullet, which will also reduce accuracy. Bullets must have a surface that forms this seal without excessive friction. These interactions between bullet and bore are termed internal ballistics. Bullets must be produced to a high standard, as surface imperfections can affect firing accuracy.
The physics affecting the bullet once it leaves the barrel is termed external ballistics. The primary factors affecting the aerodynamics of a bullet in flight are the bullet's shape and the rotation imparted by the rifling of the gun barrel. Rotational forces stabilize the bullet gyroscopically as well as aerodynamically. Any asymmetry in the bullet is largely canceled as it spins. However, a spin rate greater than the optimum value adds more trouble than good, by magnifying the smaller asymmetries or sometimes resulting in the bullet exploding midway in flight. With smooth-bore firearms, a spherical shape was optimum because no matter how it was oriented, it presented a uniform front. These unstable bullets tumbled erratically and provided only moderate accuracy; however, the aerodynamic shape changed little for centuries. Generally, bullet shapes are a compromise between aerodynamics, interior ballistic necessities, and terminal ballistics requirements. Another method of stabilization is to place the center of mass of the bullet as far forward as is practical, which is how the Minié ball and the shuttlecock are designed. This makes the bullet fly front-forward by means of aerodynamics.
Terminal ballistics and stopping power are aspects of bullet design that affect what happens when a bullet impacts with an object. The outcome of the impact is determined by the composition and density of the target material, the angle of incidence, and the velocity and physical characteristics of the bullet itself. Bullets are generally designed to penetrate, deform, or break apart. For a given material and bullet, the strike velocity is the primary factor that determines which outcome is achieved.
Bullet shapes are many and varied, and an array of them can be found in any reloading manual that sells bullet molds. Mold manufacturers such as RCBS, Paul Jones Moulds, and David Mos offer many different calibers and designs. With a mold, bullets can be made at home for reloading ammunition, where local laws allow. Hand-casting, however, is only time- and cost-effective for solid lead bullets. Cast and jacketed bullets are also commercially available from numerous manufacturers for hand loading and are most often
more convenient than casting bullets from bulk or scrap lead.
Propulsion
Propulsion of the ball can happen via several methods:
by using only gunpowder (i.e. as in flintlock weapons)
by using a percussion cap and gunpowder (i.e. as in percussion weapons)
by using a cartridge (which contains primer, gunpowder and bullet in a single package)
Materials
Bullets for black powder, or muzzle-loading firearms, were classically molded from pure lead. This worked well for low-speed bullets, fired at velocities of less than 450 m/s (1,475 ft/s). For slightly higher-speed bullets fired in modern firearms, a harder alloy of lead and tin or typesetter's lead (used to mold linotype) works very well. For even higher-speed bullet use, jacketed coated lead bullets are used. The common element in all of these, lead, is widely used because it is very dense, thereby providing a high amount of mass—and thus, kinetic energy—for a given volume. Lead is also cheap, easy to obtain, easy to work, and melts at a low temperature, which results in comparatively easy fabrication of bullets.
Lead: Simple cast, extruded, swaged, or otherwise fabricated lead slugs are the simplest form of bullets. At speeds of greater than 300 m/s (1,000 ft/s) (common in most handguns), lead is deposited in rifled bores at an ever-increasing rate. Alloying the lead with a small percentage of tin and/or antimony serves to reduce this effect, but grows less effective as velocities are increased. A cup made of harder metal, such as copper, placed at the base of the bullet and called a gas check, is often used to decrease lead deposits by protecting the rear of the bullet against melting when fired at higher pressures, but this too does not solve the problem at higher velocities. A modern solution is to powder coat the lead projectile, encasing it in a protective skin, allowing higher velocities to be achieved without lead deposits.
Jacketed lead: Bullets intended for even higher-velocity applications generally have a lead core that is jacketed or plated with gilding metal, cupronickel, copper alloys, or steel; a thin layer of harder metal protects the softer lead core when the bullet is passing through the barrel and during flight, which allows delivering the bullet intact to the target. There, the heavy lead core delivers its kinetic energy to the target. Full metal jacket or "ball" bullets (cartridges with ball bullets, which despite the name are not spherical, are called ball ammunition) are completely encased in the harder metal jacket, except for the base. Some bullet jackets do not extend to the front of the bullet, to aid expansion and increase lethality; these are called soft point (if the exposed lead tip is solid) or hollow point bullets (if a cavity or hole is present). Steel bullets are often plated with copper or other metals for corrosion resistance during long periods of storage. Synthetic jacket materials such as nylon and Teflon have been used, with limited success, especially in rifles; however, hollow point bullets with plastic aerodynamic tips have been very successful at both improving accuracy and enhancing expansion. Newer plastic coatings for handgun bullets, such as Teflon-coated bullets, are making their way into the market.
Solid or monolithic solid: Mono-metal bullets intended for deep penetration in big game animals and slender shaped very-low-drag projectiles for long range shooting are produced out of metals like oxygen free copper and alloys like copper nickel, tellurium copper and brass, for example highly machinable UNS C36000 free-cutting brass. Often these projectiles are turned on precision CNC lathes. In the case of solids, and the ruggedness of the game animals on which they are used, e.g., the African buffalo or elephant, expansion is almost entirely relinquished for the necessary penetration. In shotgunning, "slug" loads are often solid large single lead projectiles, sometimes with a hollow point, used for deer or wild pig hunting in jurisdictions that do not allow rifles (because a missed slug shot will travel considerably less far than a rifle bullet).
Fluted: In appearance, these are solid bullets with scalloped sides (missing material). The theory is that the flutes produce hydraulic jetting when passing through tissues, creating a wound channel larger than that made by conventional expanding ammunition such as hollowpoints.
Hard cast: A hard lead alloy intended to reduce fouling of rifling grooves (especially of the polygonal rifling used in some popular pistols). Benefits include simpler manufacture than jacketed bullets and good performance against hard targets; limitations are an inability to mushroom and subsequent over-penetration of soft targets.
Blank: Wax, paper, plastic, and other materials are used to simulate live gunfire and are intended only to hold the powder in a blank cartridge and to produce noise, flame and smoke. The "bullet" may be captured in a purpose-designed device or it may be allowed to expend what little energy it has in the air. Some blank cartridges are crimped or closed at the end and do not contain any bullet; some are fully loaded cartridges (without bullets) designed to propel rifle grenades. Blank cartridges, at short ranges, can be lethal due to the force of the expanding gas – numerous tragic accidents have occurred with blank cartridges (e.g., the death of actor Jon-Erik Hexum).
Practice: Made from lightweight materials like rubber, wax, wood, plastic, or lightweight metal, practice bullets are intended for short-range target work, only. Because of their weight and low velocity, they have limited ranges.
Polymer: These are metal-polymer composites, generally lighter and having higher velocities than pure metal bullets of the same dimensions. They permit unusual designs that are difficult with conventional casting or lathing.
Less lethal, or less than lethal: Rubber bullets, plastic bullets, and beanbags are designed to be non-lethal, for example for use in riot control. They are generally low velocity and are fired from shotguns, grenade launchers, paint ball guns, or specially designed firearms and air gun devices.
Incendiary: These bullets are made with explosive or flammable mixtures in the tips that are designed to ignite on contact with a target. The intent is to ignite fuel or munitions in the target area, thereby adding to the destructive power of the bullet itself.
Exploding: Similar to the incendiary bullet, this type of projectile is designed to explode upon hitting a hard surface, preferably the bone of the intended target. Not to be mistaken for cannon shells or grenades with fuse devices, these bullets have only cavities filled with a small amount of low explosive depending on the velocity and deformation upon impact to detonate. Exploding bullets have been used in various aircraft machine guns and in anti-materiel rifles.
Tracer: These have hollow backs, filled with a flare material. Usually this is a mixture of magnesium, a perchlorate, and strontium salts to yield a bright red color, although other materials providing other colors have also sometimes been used. Tracer material burns out after a certain amount of time. Such ammunition is useful to the shooter as a means of learning how to point shoot moving targets with rifles. This type of round is also used by all branches of the United States military in combat environments as a signaling device to friendly forces. Normally it is loaded at a four to one ratio with ball ammunition and is intended to show where the shooter are firing so friendly forces can engage the target as well. The flight characteristics of tracer rounds differ from normal bullets due to their lighter weight.
Armor-piercing: Jacketed designs where the core material is a very hard, high-density metal such as tungsten, tungsten carbide, depleted uranium, or steel. A pointed tip is often used, but a flat tip on the penetrator portion is generally more effective.
Nontoxic shot: Steel, bismuth, tungsten, and other exotic bullet alloys prevent release of toxic lead into the environment. Regulations in several countries mandate the use of nontoxic projectiles especially when hunting waterfowl. It has been found that birds swallow small lead shot for their gizzards to grind food (as they would swallow pebbles of similar size), and the effects of lead poisoning by constant grinding of lead pellets against food means lead poisoning effects are magnified. Such concerns apply primarily to shotguns, firing pellets (shot) and not bullets, but there is evidence suggesting that consumption of spent rifle and pistol ammunition is also hazardous to wildlife. Reduction of hazardous substances (RoHS) legislation has also been applied to bullets on occasion to reduce the impact of lead on the environment at shooting ranges. The United States Environmental Protection Agency announced that the agency does not have the legal authority to regulate this type of product (lead bullets) under the Toxic Substances Control Act (TSCA), nor is the agency seeking such authority. With some nontoxic shot, e.g., steel shot, care must be taken to shoot only in shotguns (and with chokes) specifically designed and designated for steel shot; for other, particularly older, shotguns, serious damage to the barrel and chokes can occur. And, because steel is lighter and less dense than lead, larger sized pellets must be used, thus reducing the number of pellets in a given charge of shot and possibly limiting patterns on the target; other formulations, e.g. bismuth, do not present this disability.
Blended-metal: Bullets made using cores from powdered metals other than lead with binder or sometimes sintered.
Frangible: Designed to disintegrate into tiny particles upon impact to minimize their penetration for reasons of range safety, to limit environmental impact, or to limit the shoot-through danger behind the intended target. An example is the Glaser Safety Slug, usually a pistol caliber bullet made from an amalgam of lead shot and a hard (and thus frangible) plastic binder designed to penetrate a human target and release its component shot pellets without exiting the target.
Multiple impact bullet: Bullets that are made of separate slugs that fit together inside the cartridge, and act as a single projectile inside the barrel as they are fired. The projectiles part in flight, but are held in formation by tethers that keep the individual parts of the "bullet" from flying too far away from each other. The intention of such ammo is to increase hit chance by giving a shot-like spread to rifled slug firing guns, while maintaining a consistency in shot groupings. Multiple impact bullets may be less stable in flight than conventional solid bullets because of the added aerodynamic drag from the tether line holding the pieces in formation, and each projectile affects the flight of all the others. This may limit the benefit provided by the spread of each bullet at longer ranges.
Treaties and prohibitions
Poisonous bullets were a subject to an international agreement as early as the Strasbourg Agreement (1675).
The St. Petersburg Declaration of 1868 prohibited the use of explosive projectiles weighing less than 400 grams. It was reasoned that more deadly bullets would lead to less suffering.
The Hague Convention prohibits certain kinds of ammunition for use by uniformed military personnel against the uniformed military personnel of opposing forces. These include projectiles that explode within an individual, poisoned and expanding bullets.
Protocol III of the 1983 Convention on Certain Conventional Weapons, an annexe to the Geneva Conventions, prohibits the use of incendiary ammunitions against civilians.
Some jurisdictions acting on environmental concerns have banned hunting with lead bullets and shotgun pellets.
In December 2014, a federal appeals court denied a lawsuit by environmental groups that the EPA must use the Toxic Substances Control Act to regulate lead in shells and cartridges. The groups sought to regulate "spent lead", yet EPA could not regulate spent lead without also regulating cartridges and shells, per the court.
Abbreviations
2F – Two-part controlled fragmenting
ACC – Remington Accelerator (see sabot)
ACP – Automatic Colt Pistol
AE – Action express
AGS – African Grand Slam (Speer)
AP – Armor piercing (has a depleted uranium or other hard metal core)
APT – Armor-piercing tracer
API – Armor-piercing incendiary
API-T – Armor-piercing incendiary tracer
APFSDS – Armor-piercing fin stabilized discarding sabot round
B – Ball
B2F – Brass two-part fragmenting
BBWC – Bevel base wadcutter
BEB – Brass enclosed base
BJHP – Brass jacketed hollow point
Blitz – Sierra BlitzKing
BMG – Browning Machine Gun
BrPT – Bronze point
Bt – Boat-tail
BtHP – Boat-tail hollow point
C2F – Civilian two-part fragmenting
CB – Cast bullet
CL, C-L – Remington Core-Lokt
CMJ – Complete metal jacket, electroplated, not truly jacketed
CN – Cupronicknel
CNCS – Cupronickel-clad steel
CTFB – Closed tip flat Bbse
DBBWC – Double bevel based wadcutter
DEWC – Double ended wadcutter
DGS – Dangerous game solid (Hornady)
DGX – Dangerous game expanding (Hornady)
DU – Depleted uranium
EFMJ – Expanding full metal jacket
EVO, FTX – Hornady Leverevolution flex tip expanding
EVO – RWS Evolution bullet
FMC – Full metal case
FMJ – Full metal jacket
FMJBT – Full metal jacket boat-tail
FN – Dangerous game solid flat nose
FNEB – Flat nose enclosed base
FP – Flat point
FP – Full patch
FST – Winchester Fail Safe Talon
GAP (G.A.P.) – Glock Automatic Pistol
GC – Gas check
GD – Speer Gold Dot
GDHP – Speer Gold Dot hollow point
GM – Gilding metal
GMCS – Gilding metal-clad steel
GS – Remington Golden Saber
GSC – GS Custom turned copper bullet
HAP – Hornady Action Pistol
HBWC – Hollow base wadcutter
HC – Hard cast
HE-IT – High explosive incendiary tracer
HFN – Hard cast flat nose
HP – Hollow point
HPBT – Hollow point boat tail
HPCB – Heavy plate concave base
HPJ – High performance jacketed
HS – Federal Hydra-Shok
HST – Federal Hi-Shok Two
HV – Low friction drive band high velocity
ID-Classic – RWS fragmentation bullet, ex-TIG after Brenneke-license was not renewed
I-T – Incendiary tracer
IB – Interbond (Hornady)
J – Jacketed
JAP – Jacketed aluminium point
JFP – Jacketed flat point
JHC – Jacketed hollow cavity
JHP – Jacketed hollow point
JHP/sabot – Jacketed hollow point/sabot
JSP – Jacketed soft point
L – Lead
L-C – Lead combat
L-T – Lead target
LF – Lead free
LFN – Long flat nose
LFP – Lead flat point
LHP – Lead hollow point
LRN – Lead round nose
LSWC – Lead semiwadcutter
LSWC-GC – Lead semiwadcutter gas checked
LWC – Lead wadcutter
LTC – Lead truncated cone
MC – Metal cased
MHP – Match hollow point
MK – Sierra MatchKing
MRWC – Mid-range wadcutter
MP – Metal point
NP – Nosler partition
OTM – Open tip match
OWC – Ogival wadcutter
P – Practice, proof
PB – Lead bullet
PB – Parabellum
PL – Remington Power-Lokt
PnPT – Pneumatic point
PPL – Paper patched lead
PSP – Plated soft point
PSP, PTDSP – Pointed soft point
PRN – Plated round nose
RBT – Rebated boat tail
RN – Round nose
RNFP – Round nose flat point
RNL – Round nose lead
SCHP – Solid copper hollow point
SJ – Semi-jacketed
SJHP – Semi-jacketed hollow point
SJSP – Semi-jacketed soft point
SLAP – Saboted light armor penetrator
SP – Soft point
SP – Spire point
Sp, SPTZ – Spitzer
SPC – Special purpose cartridge
SpHP – Spitzer hollow point
SST – Hornady Super shock tip
SSp – Semi-spitzer
ST – Silver tip
STHP – Silver tip hollow point
SWC – Semiwadcutter
SX – Super explosive
SXT – Winchester Ranger supreme expansion technology
T – Tracer
TAG – Brenneke lead-free bullet ()
TBBC – Carter/Speer Trophy Bonded Bear Claw soft point
TBSS – Carter/Speer Trophy Bonded Sledgehammer solid
TC – Truncated cone
THV – Terminal high velocity
TIG – Brenneke fragmentation bullet ()
TMJ – Total metal jacket
TNT – Speer TNT
TUG – Brenneke deformation bullet ()
TOG – Brenneke deformation bullet ()
UmbPT – Umbrella point
UNI-Classic – RWS deformation bullet, ex-TUG after Brenneke-license was not renewed.
VMAX – Hornady V-Max
VLD – Very low drag
WC – Wadcutter
WFN – Wide flat nose
WFNGC – Wide flat nose gas check
WLN – Wide long nose
X – Barnes X-bullet
XTP – Hornady Extreme Terminal Performance
See also
Bullet bow shockwave
Flechette
Grain (unit)
List of handgun cartridges
List of rifle cartridges
Meplat
Multiple projectile ammunition
Shotgun shells
Smart bullet
Table of handgun and rifle cartridges
References
External links
High speed imaging of in flight bullet transition ballistics
"Bullets by the Billion", 1946 – Story about the manufacture of small caliber ammunition during World War II
Arizona Gun List – ammunition types
Bullets for Beginners
Dangerous Game Bullets
European Ammunition Box Translations
Remington Core-Lokt, Bronze Point & Power-Lokt Centerfire Ammunition
How To Make GUN BULLET | How To Machines
Projectiles
Articles containing video clips |
null | null | Space tourism | eng_Latn | Space tourism is human space travel for recreational purposes. There are several different types of space tourism, including orbital, suborbital and lunar space tourism.
During the period from 2001 to 2009, seven space tourists made eight space flights aboard a Russian Soyuz spacecraft to the International Space Station, brokered by Space Adventures in conjunction with Roscosmos and RSC Energia. The publicized price was in the range of US$20–25 million per trip. Some space tourists have signed contracts with third parties to conduct certain research activities while in orbit. By 2007, space tourism was thought to be one of the earliest markets that would emerge for commercial spaceflight.
Russia halted orbital space tourism in 2010 due to the increase in the International Space Station crew size, using the seats for expedition crews that would previously have been sold to paying spaceflight participants. Orbital tourist flights were set to resume in 2015 but the planned flight was postponed indefinitely. Russian orbital tourism eventually resumed with the launch of Soyuz MS-20 in 2021.
On June 7, 2019, NASA announced that starting in 2020, the organization aims to start allowing private astronauts to go on the International Space Station, with the use of the SpaceX Crew Dragon spacecraft and the Boeing Starliner spacecraft for public astronauts, which is planned to be priced at 35,000 USD per day for one astronaut, and an estimated 50 million USD for the ride there and back.
Work also continues towards developing suborbital space tourism vehicles. This is being done by aerospace companies like Blue Origin and Virgin Galactic. SpaceX announced in 2018 that they are planning on sending space tourists, including Yusaku Maezawa, on a free-return trajectory around the Moon on the Starship.
Precursors
The Soviet space program was successful in broadening the pool of cosmonauts. The Soviet Intercosmos program included cosmonauts selected from Warsaw Pact member countries (Czechoslovakia, Poland, East Germany, Bulgaria, Hungary, Romania) and later from allies of the USSR (Cuba, Mongolia, Vietnam) and non-aligned countries (India, Syria, Afghanistan). Most of these cosmonauts received full training for their missions and were treated as equals, but were generally given shorter flights than Soviet cosmonauts. The European Space Agency (ESA) also took advantage of the program.
The US space shuttle program included payload specialist positions which were usually filled by representatives of companies or institutions managing a specific payload on that mission. These payload specialists did not receive the same training as professional NASA astronauts and were not employed by NASA. In 1983, Ulf Merbold from the ESA and Byron Lichtenberg from MIT (engineer and Air Force fighter pilot) were the first payload specialists to fly on the Space Shuttle, on mission STS-9.
In 1984, Charles D. Walker became the first non-government astronaut to fly, with his employer McDonnell Douglas paying US$40,000 () for his flight. During the 1970s, Shuttle prime contractor Rockwell International studied a $200–300 million removable cabin that could fit into the Shuttle's cargo bay. The cabin could carry up to 74 passengers into orbit for up to three days. Space Habitation Design Associates proposed, in 1983, a cabin for 72 passengers in the bay. Passengers were located in six sections, each with windows and its own loading ramp, and with seats in different configurations for launch and landing. Another proposal was based on the Spacelab habitation modules, which provided 32 seats in the payload bay in addition to those in the cockpit area. A 1985 presentation to the National Space Society stated that, although flying tourists in the cabin would cost $1 million to $1.5 million per passenger without government subsidy, within 15 years, 30,000 people a year would pay US$25,000 () each to fly in space on new spacecraft. The presentation also forecast flights to lunar orbit within 30 years and visits to the lunar surface within 50 years.
As the shuttle program expanded in the early 1980s, NASA began a Space Flight Participant program to allow citizens without scientific or governmental roles to fly. Christa McAuliffe was chosen as the first Teacher in Space in July 1985 from 11,400 applicants. 1,700 applied for the Journalist in Space program. An Artist in Space program was considered, and NASA expected that after McAuliffe's flight two to three civilians a year would fly on the shuttle. After McAuliffe was killed in the Challenger disaster in January 1986, the programs were canceled. McAuliffe's backup, Barbara Morgan, eventually got hired in 1998 as a professional astronaut and flew on STS-118 as a mission specialist. A second journalist-in-space program, in which NASA green-lighted Miles O'Brien to fly on the space shuttle, was scheduled to be announced in 2003. That program was canceled in the wake of the Columbia disaster on STS-107 and subsequent emphasis on finishing the International Space Station before retiring the Space Shuttle.
Initially, senior figures at NASA strongly opposed space tourism on principle; from the beginning of the ISS expeditions, NASA stated it was not interested in accommodating paying guests. The Subcommittee on Space and Aeronautics Committee on Science of the House of Representatives held in June 2001 revealed the shifting attitude of NASA towards paying space tourists wanting to travel to the ISS in its statement on the hearing's purpose: "Review the issues and opportunities for flying nonprofessional astronauts in space, the appropriate government role for supporting the nascent space tourism industry, use of the Shuttle and Space Station for Tourism, safety and training criteria for space tourists, and the potential commercial market for space tourism." The subcommittee report was interested in evaluating Dennis Tito's extensive training and his experience in space as a nonprofessional astronaut.
With the realities of the post-Perestroika economy in Russia, its space industry was especially starved for cash. The Tokyo Broadcasting System (TBS) offered to pay for one of its reporters to fly on a mission. Toyohiro Akiyama was flown in 1990 to Mir with the eighth crew and returned a week later with the seventh crew. Cost estimates vary from $10 million up to $37 million. Akiyama gave a daily TV broadcast from orbit and also performed scientific experiments for Russian and Japanese companies.
In 1991, British chemist Helen Sharman was selected from a pool of 13,000 applicants to be the first Briton in space. The program was known as Project Juno and was a cooperative arrangement between the Soviet Union and a group of British companies. The Project Juno consortium failed to raise the funds required, and the program was almost canceled. Reportedly Mikhail Gorbachev ordered it to proceed under Soviet expense in the interests of international relations, but in the absence of Western underwriting, less expensive experiments were substituted for those in the original plans. Sharman flew aboard Soyuz TM-12 to Mir and returned aboard Soyuz TM-11.
Sub-orbital space tourism
Successful projects
Scaled Composites won the $10 million X Prize in October 2004 with SpaceShipOne, as the first private company to reach and surpass an altitude of twice within two weeks. The altitude is beyond the Kármán Line, the arbitrarily defined boundary of space. The first flight was flown by Michael Melvill in June 2004, to a height of , making him the first commercial astronaut. The prize-winning flight was flown by Brian Binnie, which reached a height of , breaking the X-15 record. There were no space tourists on the flights even though the vehicle has seats for three passengers. Instead there was additional weight to make up for the weight of passengers.
In 2005, Virgin Galactic was founded as a joint venture between Scaled Composites and Richard Branson's Virgin Group. Eventually Virgin Group owned the entire project. Virgin Galactic began building SpaceShipTwo-class spaceplanes. The first of these spaceplanes, VSS Enterprise, was intended to commence its first commercial flights in 2015, and tickets were on sale at a price of $200,000 (later raised to $250,000). However, the company suffered a considerable setback when the Enterprise broke up over the Mojave Desert during a test flight in October 2014. Over 700 tickets had been sold prior to the accident. A second spaceplane, VSS Unity, completed a successful test flight with four passengers on July 11, 2021.
Blue Origin developed the New Shepard reusable suborbital launch system specifically to enable short-duration space tourism. Blue Origin plans to ferry a maximum of six persons on a brief journey to space on board the New Shepard. The capsule is attached to the top portion of an 18-meter (59-foot) rocket. The rocket successfully launched with four passengers on July 20, 2021, and reached an altitude of .
Canceled projects
Armadillo Aerospace was developing a two-seat vertical takeoff and landing (VTOL) rocket called Hyperion, which will be marketed by Space Adventures. Hyperion uses a capsule similar in shape to the Gemini capsule. The vehicle will use a parachute for descent but will probably use retrorockets for final touchdown, according to remarks made by Armadillo Aerospace at the Next Generation Suborbital Researchers Conference in February 2012. The assets of Armadillo Aerospace were sold to Exos Aerospace and while SARGE is continuing to be developed, it is unclear whether Hyperion is still being developed.
XCOR Aerospace was developing a suborbital vehicle called Lynx until development was halted in May 2016. The Lynx would take off from a runway under rocket power. Unlike SpaceShipOne and SpaceShipTwo, Lynx would not require a mothership. Lynx was designed for rapid turnaround, which would enable it to fly up to four times per day. Because of this rapid flight rate, Lynx had fewer seats than SpaceShipTwo, carrying only one pilot and one spaceflight participant on each flight. XCOR expected to roll out the first Lynx prototype and begin flight tests in 2015, but as of late 2017, XCOR was unable to complete their prototype development and filed for bankruptcy.
Citizens in Space, formerly the Teacher in Space Project, is a project of the United States Rocket Academy. Citizens in Space combines citizen science with citizen space exploration. The goal is to fly citizen-science experiments and citizen explorers (who travel free) who will act as payload operators on suborbital space missions. By 2012, Citizens in Space had acquired a contract for 10 suborbital flights with XCOR Aerospace and expected to acquire additional flights from XCOR and other suborbital spaceflight providers in the future. In 2012, Citizens in Space reported they had begun training three citizen astronaut candidates and would select seven additional candidates over the next 12 to 14 months.
Space Expedition Corporation was preparing to use the Lynx for "Space Expedition Curaçao", a commercial flight from Hato Airport on Curaçao, and planned to start commercial flights in 2014. The costs were $95,000 each.
Axe Apollo Space Academy promotion by Unilever which planned to provide 23 people suborbital spaceflights on board the Lynx.
EADS Astrium, a subsidiary of European aerospace giant EADS, announced its space tourism project in June 2007.
Orbital space tourism
As of 2021, Space Adventures and SpaceX are the only companies to have coordinated tourism flights to Earth's orbit. Virginia-based Space Adventures has worked with Russia to use its Soyuz spacecraft to fly ultra-wealthy individuals to the International Space Station. The tourists included entrepreneur and space investor Anousheh Ansari and Cirque du Soleil co-founder Guy Laliberté. Those missions were priced at around $20 million each. The space industry could soon be headed for a tourism revolution if SpaceX and Boeing make good on their plans to take tourists to orbit.
Successful projects
At the end of the 1990s, MirCorp, a private venture that was by then in charge of the space station, began seeking potential space tourists to visit Mir in order to offset some of its maintenance costs. Dennis Tito, an American businessman and former JPL scientist, became their first candidate. When the decision was made to de-orbit Mir, Tito managed to switch his trip to the International Space Station (ISS) aboard a Russian Soyuz spacecraft through a deal between MirCorp and US-based Space Adventures, Ltd. Dennis Tito visited the ISS for seven days in April–May 2001, becoming the world's first "fee-paying" space tourist. Tito paid a reported $20 million for his trip.
Tito was followed in April 2002 by South African Mark Shuttleworth (Soyuz TM-34). The third was Gregory Olsen in October 2005 (Soyuz TMA-7). In February 2003, the Space Shuttle Columbia disintegrated on re-entry into the Earth's atmosphere, killing all seven astronauts aboard. After this disaster, space tourism on the Russian Soyuz program was temporarily put on hold, because Soyuz vehicles became the only available transport to the ISS. After the Shuttle return to service in July 2005, space tourism was resumed. In September 2006, an Iranian American businesswoman named Anousheh Ansari became the fourth space tourist (Soyuz TMA-9).) In April 2007, Charles Simonyi, an American businessman of Hungarian descent, joined their ranks (Soyuz TMA-10). Simonyi became the first repeat space tourist, paying again to fly on Soyuz TMA-14 in March 2009. British-American Richard Garriott became the next space tourist in October 2008 aboard Soyuz TMA-13. Canadian Guy Laliberté visited the ISS in September 2009 aboard Soyuz TMA-16, becoming the last visiting tourist until December 2021. Originally the third member aboard Soyuz TMA-18M should have been the British singer Sarah Brightman as a space tourist, but on May 13, 2015, she announced she had withdrawn from training.
Since the Space Shuttle was retired in 2011, Soyuz once again became the only means of accessing the ISS, and so tourism was once again put on hold. On June 7, 2019, NASA announced a plan to open the ISS to space tourism again.
On September 16, 2021, the Inspiration4 mission launched from the Kennedy Space Center on a SpaceX Falcon 9 and spent almost three days in orbit aboard the Crew Dragon Resilience, becoming the first all-civilian crew to fly an orbital space mission.
Ongoing projects
The Boeing Starliner capsule is being developed as part of the NASA's Commercial Crew Program. Part of the agreement with NASA allows Boeing to sell seats for space tourists. Boeing proposed including one seat per flight for a spaceflight participant at a price that would be competitive with what Roscosmos charges tourists.
SpaceX Ax-1: Axiom Space and SpaceX plan to send tourists to the ISS using a Dragon 2 spacecraft. The flight is targeted for launch no earlier than the end of March 2022.
Canceled projects
In 2004, Bigelow Aerospace established a competition called America's Space Prize, which offered a $50 million prize to the first US company to create a reusable spacecraft capable of carrying passengers to a Nautilus space station. The prize expired in January 2010 without anyone making a serious effort to win it.
The Space Island Group proposed having 20,000 people on their "space island" by 2020.
A United States startup firm, Orion Span announced during the early part of 2018 that it planned to launch and position a luxury space hotel in orbit within several years. Aurora Space Station, the name of the hotel, would have offered guests (at most six individuals) 12 days of staying in a pill-shaped space hotel for $9.5 million. The hotel's cabins would have measured approximately 43 feet by 14 feet in width.
Space Adventures Crew Dragon mission: Space Adventures and SpaceX planned to send up to four tourists to low Earth orbit for a few days in late 2021 or early 2022. In October 2021, Space Adventures stated that the mission contract had expired, though the possibility of a future partnership with SpaceX was left open.
Galactic Suite Design
Orbital Technologies Commercial Space Station
Space Industries Incorporated
Space Islands
Tourism beyond Earth orbit
Ongoing projects
In February 2017, Elon Musk announced that substantial deposits from two individuals had been received by SpaceX for a Moon loop flight using a free return trajectory and that this could happen as soon as late 2018. Musk said that the cost of the mission would be "comparable" to that of sending an astronaut to the International Space Station, about US$70 million in 2017. In February 2018, Musk announced that the Falcon Heavy rocket would not be used for crewed missions. The proposal changed in 2018 to use the Starship launch system instead. In September 2018, Musk revealed the passenger for the trip, Yusaku Maezawa during a livestream. Yusaku Maezawa described the plan for his trip in further detail, dubbed the #dearMoon project, intending to take 6–8 artists with him on the journey to inspire the artists to create new art.
Space Adventures Ltd. have announced that they are working on DSE-Alpha, a circumlunar mission to the Moon, with the price per passenger being $100,000,000.
Canceled projects
Excalibur Almaz proposed to take three tourists in a flyby around the Moon, using modified Almaz space station modules, in a low-energy trajectory flyby around the Moon. The trip would last around 6 months. However, their equipment was never launched and is to be converted into an educational exhibit.
The Golden Spike Company was an American space transport startup active from 2010 to 2013. The company held the objective to offer private commercial space transportation services to the surface of the Moon. The company's website was quietly taken offline in September 2015.
The Inspiration Mars Foundation is an American nonprofit organization founded by Dennis Tito that proposed to launch a crewed mission to flyby Mars in January 2018, or 2021 if they missed the first deadline. Their website became defunct by late 2015 but it is archived by the Internet Archive. The Foundation's future plans are unclear.
Bigelow Aerospace planned to extend their successes with the Genesis modules by launching the B330, an expandable habitation module with 330 cubic meters of internal space, aboard a Vulcan rocket. The Vulcan was contracted to boost BA 330 to low lunar orbit by the end of 2022.
Legality
Under the Outer Space Treaty signed in 1967, the launch operator's nationality and the launch site's location determine which country is responsible for any damages occurred from a launch.
After valuable resources were detected on the Moon, private companies began to formulate methods to extract the resources. Article II of the Outer Space Treaty dictates that "outer space, including the Moon and other celestial bodies, is not subject to national appropriation by claim of sovereignty, by means of use or occupation, or by any other means". However, countries have the right to freely explore the Moon and any resources collected are property of that country when they return.
United States
In December 2005, the US government released a set of proposed rules for space tourism. These included screening procedures and training for emergency situations, but not health requirements.
Under current US law, any company proposing to launch paying passengers from American soil on a suborbital rocket must receive a license from the Federal Aviation Administration's Office of Commercial Space Transportation (FAA/AST). The licensing process focuses on public safety and safety of property, and the details can be found in the Code of Federal Regulations, Title 14, Chapter III. This is in accordance with the Commercial Space Launch Amendments Act passed by Congress in 2004.
In March 2010, the New Mexico legislature passed the Spaceflight Informed Consent Act. The SICA gives legal protection to companies who provide private space flights in the case of accidental harm or death to individuals. Participants sign an Informed Consent waiver, dictating that spaceflight operators cannot be held liable in the "death of a participant resulting from the inherent risks of space flight activities". Operators are however not covered in the case of gross negligence or willful misconduct.
List of space tourism trips
The following list notes each trip taken by an individual for whom a fee was paid (by themselves or another party) to go above the Kármán Line, the internationally recognized boundary of space. It also includes future trips which are paid for and scheduled.
Criticism of the term space tourist
Many private space travelers have objected to the term space tourist, often pointing out that their role went beyond that of an observer, since they also carried out scientific experiments in the course of their journey. Richard Garriott additionally emphasized that his training was identical to the requirements of non-Russian Soyuz crew members, and that teachers and other non-professional astronauts chosen to fly with NASA are called astronauts. He has said that if the distinction has to be made, he would rather be called "private astronaut" than "tourist". Mark Shuttleworth described himself as a "pioneer of commercial space travel". Gregory Olsen prefers "private researcher", and Anousheh Ansari prefers the term "private space explorer". Other space enthusiasts object to the term on similar grounds. Rick Tumlinson of the Space Frontier Foundation, for example, has said: "I hate the word tourist, and I always will ... 'Tourist' is somebody in a flowered shirt with three cameras around his neck." Russian cosmonaut Maksim Surayev told the press in 2009 not to describe Guy Laliberté as a tourist: "It's become fashionable to speak of space tourists. He is not a tourist but a participant in the mission."
"Spaceflight participant" is the official term used by NASA and the Russian Federal Space Agency to distinguish between private space travelers and career astronauts. Tito, Shuttleworth, Olsen, Ansari, and Simonyi were designated as such during their respective space flights. NASA also lists Christa McAuliffe as a spaceflight participant (although she did not pay a fee), apparently due to her non-technical duties aboard the STS-51-L flight.
The US Federal Aviation Administration awards the title of "commercial astronaut" to trained crew members of privately funded spacecraft. The only people currently holding this title are Mike Melvill and Brian Binnie, the pilots of SpaceShipOne in 2004; pilots Mark P. Stucky and Frederick W. Sturckow in 2018, and pilots Dave Mackay, Michael Masucci, and trainer Beth Moses in 2019 aboard SpaceShipTwo on two separate missions.
Attitudes towards space tourism
A 2018 survey from the PEW Research Center identifies the top three motivations for a customer to purchase a flight into space as:
To experience something unique ( e.g. pioneering, one of a kind)
To see the view of Earth from space
To learn more about the world
The PEW study also found that only 43% of Americans would be definitely or probably interested in going into space.
A web-based survey suggested that over 70% of those surveyed wanted less than or equal to 2 weeks in space; in addition, 88% wanted to spacewalk, of whom 14% would pay a 50% premium for the experience, and 21% wanted a hotel or space station.
The concept has met with some criticism; Günter Verheugen, vice-president of the European Commission, said of the EADS Astrium Space Tourism Project: "It's only for the super-rich, which is against my social convictions".
On 14 October 2021, Prince William suggested that entrepreneurs should focus on saving Earth rather than engaging in space tourism and also warned about a rise in "climate anxiety" among younger generations.
Environmental effects
A 2010 study published in Geophysical Research Letters raised concerns that the growing commercial spaceflight industry could accelerate global warming. The study, funded by NASA and The Aerospace Corporation, simulated the impact of 1,000 suborbital launches of hybrid rockets from a single location, calculating that this would release a total of 600 tonnes of black carbon into the stratosphere. They found that the resultant layer of soot particles remained relatively localized, with only 20% of the carbon straying into the southern hemisphere, thus creating a strong hemispherical asymmetry. This unbalance would cause the temperature to decrease by about in the tropics and subtropics, whereas the temperature at the poles would increase by between . The ozone layer would also be affected, with the tropics losing up to 1.7% of ozone cover, and the polar regions gaining 5–6%. The researchers stressed that these results should not be taken as "a precise forecast of the climate response to a specific launch rate of a specific rocket type", but as a demonstration of the sensitivity of the atmosphere to the large-scale disruption that commercial space tourism could bring.
Education and advocacy
Several organizations have been formed to promote the space tourism industry, including the Space Tourism Society, Space Future, and HobbySpace. UniGalactic Space Travel Magazine is a bi-monthly educational publication covering space tourism and space exploration developments in companies like SpaceX, Orbital Sciences, Virgin Galactic and organizations like NASA.
Classes in space tourism are currently taught at the Rochester Institute of Technology in New York, and Keio University in Japan.
Economic potential
A 2010 report from the Federal Aviation Administration, titled "The Economic Impact of Commercial Space Transportation on the U.S. Economy in 2009", cites studies done by Futron, an aerospace and technology-consulting firm, which predict that space tourism could become a billion-dollar market within 20 years. Eight tourists reached orbit between 2001 and 2009. In 2011 Space Adventures suggested that this number could reach 140 by 2020, but with commercial crewed rockets only just beginning to enter service, such numbers have yet to be achieved.
See also
Commercialization of space
Effect of spaceflight on the human body
Private spaceflight
Space flight participant
Sub-orbital spaceflight
Commercial astronaut
References
Further reading
External links
Space Tourists, a documentary film by Christian Frei
Space Tourism Society, a non-profit research and advocacy group
American inventions
Russian inventions
Types of tourism
2001 introductions |
null | null | Fred Dagg | eng_Latn | Fred Dagg is a fictional character from New Zealand created and acted on stage, film and television by satirist John Clarke. Clarke graced New Zealand TV screens as Dagg during the mid to late 1970s, "taking the piss" out of the post-pioneering Kiwi bloke and "blokesses". The sense of dag being also associated with the term dag.
The Fred Dagg character is a stereotypical farmer and New Zealand bloke: clad in a black singlet and gumboots, hailing from the isolated rural town of Taihape, and attended by numerous associates (or sons) all named "Trev". One memorable expression was uttered whenever there was a knock at the door: "That'll be the door."
When Clarke first unveiled the character of Fred Dagg in recordings and on New Zealand TV in 1975, he became a national star. Clarke also recorded a series of records and cassettes as Dagg, as well as publishing several books.
In 1977, Clarke moved from New Zealand to Australia, where he went on to establish himself as a top script writer and personality. He appeared regularly on Australian television delivering political satire sketches with actor comedian Bryan Dawe until his death in April 2017.
Influence
Clarke has stated that the inspiration for Fred Dagg came primarily from his uncles who would talk in such a way, using intonation and speech rhythm, to cause laughter without the standard telling of jokes.
Much of the comedy coming into New Zealand at the time was British radio and this was an inspiration to him.
Works
John Clarke slowly refined the character of Fred Dagg in short comedic interview segments that screened occasionally on the New Zealand current affairs show Nationwide throughout 1974. In December of that year the character was the subject of a satirical mockumentary spoof episode on the NZBC farming show, Country Calendar. Fred Dagg went on to become a household name in New Zealand in 1975 with further comedy skit appearances on the Television One news programme Tonight at Nine and with the release by Clarke (as Dagg) of two singles with EMI, "Traditional Air"/"Unlabelled", and "We Don't Know How Lucky We Are"/"Larry Loves Barry", with the latter making it to number 17 on the national music single charts.
An album called Fred Dagg's Greatest Hits followed and was a massive seller. Thirty years after its release this album remains one of New Zealand's all-time biggest selling records.
Another single recorded with Diamond Lil was an even bigger hit in 1976. "Gumboots"/"Save The Last Dance For Me" climbed to number 6 on the charts. "Gumboots" was a modified version of Billy Connolly's "If It Wasna For Your Wellies", itself an adaptation of the old song "The Work Of The Weavers".
A second album, Fred Dagg Live was released in 1976. Following on in the style of the first, it was also a huge seller.
1977 saw the release of the film Dagg Day Afternoon, co-directed and co-written by John Clarke and Geoff Murphy, and starring Fred Dagg. The film, under 45 minutes long, is essentially a series of sketches tied together with a loose narrative about Dagg's secret mission to find a "bionic sheep" (or 6 million dollar ram) which has been lost by the government.
By 1978 John Clarke and his young family had settled in Australia but he briefly commuted back to New Zealand during that year to create a Fred Dagg television series for South Pacific Television (TV2). The Fred Dagg Lectures on Leisure consisted of 20 five-minute long episodes which screened Sunday nights from September 1978 through to February 1979. The programmes featured Fred holding forth on a variety of diverse subjects such as golf, photography, how to write an autobiography, UFO spotting, saving whales and tree-felling. The original master tapes of all these episodes no longer exist and are thought to have either been wiped or lost by SPTV when the company was dissolved and amalgamated with Television One in 1980.
A third and final album was released in 1979 called The Fred Dagg Tapes.
In 1998 the Fred Dagg Anthology CD was released by Columbia. It contained most of Dagg's best-known material. A lot of the second half of the Anthology isn't genuine "Fred Dagg" material, having been created by Clarke after he'd retired Fred Dagg and moved to Australia. A number of mock interviews were first broadcast on Australian current affairs television, and lampoon prominent people (such as businessman Alan Bond) who would not be familiar to a New Zealand audience. There are also six "farnarkeling" reports, which parody sports news and were first performed by Clarke on the ABC's The Gillies Report.
"We Don't Know How Lucky We Are!" was re-released in 1998 with revised lyrics. This re-recording was driven by Graeme Hill, and Sky TV's SportsCafe.
Discography
Studio and live albums
Compilation albums
See also
Agriculture in New Zealand
Gumboot Day
References
External links
Fred Dagg on John Clarke's official website
Fred Dagg on New Zealand Music.
Fred Dagg's gumboots at the collection of the Museum of New Zealand Te Papa Tongarewa
New Zealand humour
Fictional New Zealand people
Fictional farmers |
null | null | Mekong | eng_Latn | The Mekong or Mekong River is a trans-boundary river in East Asia and Southeast Asia. It is the world's twelfth longest river and the third longest in Asia. Its estimated length is , and it drains an area of , discharging of water annually.
From the Tibetan Plateau the river runs through China, Myanmar, Laos, Thailand, Cambodia, and Vietnam. The extreme seasonal variations in flow and the presence of rapids and waterfalls in the Mekong make navigation difficult. Even so, the river is a major trade route between western China and Southeast Asia.
Names
Mekong is originally called Mae Nam Khong from a contracted form of Tai shortened to Mae Khong. In Thai and Lao, Mae Nam ("Mother of Water[s]") is used for large rivers and Khong is the proper name referred to as "River Khong". However, Khong is an archaic word meaning "river", loaned from Austroasiatic languages, such as Vietnamese sông (from *krong) and Mon kruŋ "river", which led to Chinese whose Old Chinese pronunciation has been reconstructed as and which long served as the proper name of the Yangtze before becoming a generic word for major rivers. To the early European traders, the Mekong River was also known as Mekon River, May-Kiang River and Cambodia River.
The local names for the river include:
From Tai:
, , or just ''[mɛ̂ː kʰǒːŋ].
, , .
Tai Lue: , .
Mékôngk , Tônlé Mékôngk .
, .
or .
Chinese: .
().
Other:
, (九龍 Nine Dragons River ).
Tônlé Thum (lit. "Big River" or "Great River").
Khmuic: , '' means 'river' or 'water', here it means 'river', '' means 'canal'. So '' means 'canal river'. In the ancient time Khmuic people called it '' or '' which means 'giant canal river' or 'deep canal river' respectively.
Course
The Mekong rises as the Za Qu (; ) and soon becomes known as the Lancang (, from the old name of Lao kingdom Lan Xang; the characters may also be literally understood as "turbulent green river"). It originates in the "three rivers source area" on the Tibetan Plateau in the Sanjiangyuan National Nature Reserve. The reserve protects the headwaters of, from north to south, the Yellow (Huang He), the Yangtze, and the Mekong Rivers. It flows through the Tibetan Autonomous Region and then southeast into Yunnan Province, and then the Three Parallel Rivers Area in the Hengduan Mountains, along with the Yangtze to its east and the Salween River (Nu Jiang in Chinese) to its west.
Then the Mekong meets the China–Myanmar border and flows about along that border until it reaches the tripoint of China, Myanmar and Laos. From there it flows southwest and forms the border of Myanmar and Laos for about until it arrives at the tripoint of Myanmar, Laos, and Thailand. This is also the point of confluence between the Ruak River (which follows the Thai–Myanmar border) and the Mekong. The area of this tripoint is sometimes termed the Golden Triangle, although the term also refers to the much larger area of those three countries that was notorious as a drug producing region.
From the Golden Triangle tripoint, the Mekong turns southeast to briefly form the border of Laos with Thailand.
Khon Pi Long is a series of rapids along a 1.6-kilometre section of the Mekong River dividing Chiang Rai and Bokeo Province in Laos. The name of the rapids means 'where the ghost lost its way'. It then turns east into the interior of Laos, flowing first east and then south for some before meeting the border with Thailand again. Once more, it defines the Laos-Thailand border for some as it flows first east, passing the capital of Laos, Vientiane, then turns south. A second time, the river leaves the border and flows east into Laos soon passing the city of Pakse. Thereafter, it turns and runs more or less directly south, crossing into Cambodia.
At Phnom Penh the river is joined on the right bank by the river and lake system the Tonlé Sap. When the Mekong is low, the Tonle Sap is a tributary: water flows from the lake and river into the Mekong. When the Mekong floods, the flow reverses: the floodwaters of the Mekong flow up the Tonle Sap.
Immediately after the Sap River joins the Mekong by Phnom Penh, the Bassac River branches off the right (west) bank. The Bassac River is the first and main distributary of the Mekong. This is the beginning of the Mekong Delta. The two rivers, the Bassac to the west and the Mekong to the east, enter Vietnam shortly after this. In Vietnam, the Bassac is called the Hậu River (Sông Hậu or Hậu Giang); the main, eastern, branch of the Mekong is called the Tiền River or Tiền Giang. In Vietnam, distributaries of the eastern (main, Mekong) branch include the Mỹ Tho River, the Ba Lai River, the Hàm Luông River, and the Cổ Chiên River.
Drainage basin
The Mekong Basin can be divided into two parts: the "upper Mekong basin" in Tibet, and the "lower Mekong basin" from Yunnan downstream from China to the South China Sea. From the point where it rises to its mouth, the most precipitous drop in the Mekong occurs in the upper Mekong basin, a stretch of some . Here, it drops before it enters the lower basin where the borders of Thailand, Laos, China, and Myanmar come together in the Golden Triangle. Downstream from the Golden Triangle, the river flows for a further through Laos, Thailand, and Cambodia before entering the South China Sea via a complex delta system in Vietnam.
Upper basin
The upper basin makes up 24% of the total area and contributes 15–20% of the water that flows into the Mekong River. The catchment here is steep and narrow. Soil erosion has been a major problem and approximately 50% of the sediment in the river comes from the upper basin.
In Yunnan Province in China, the river and its tributaries are confined by narrow, deep gorges. The tributary river systems in this part of the basin are small. Only 14 have catchment areas that exceed , yet the greatest amount of loss of forest cover in the entire river system per square kilometer has occurred in this region due to heavy unchecked demand for natural resources. In the south of Yunnan, in Simao and Xishuangbanna Prefectures, the river changes as the valley opens out, the floodplain becomes wider, and the river becomes wider and slower.
Lower basin
Major tributary systems develop in the lower basin. These systems can be separated into two groups: tributaries that contribute to the major wet season flows, and tributaries that drain low relief regions of lower rainfall. The first group are left bank tributaries that drain the high rainfall areas of Laos. The second group are those on the right bank, mainly the Mun and Chi Rivers, that drain a large part of northeast Thailand.
Laos lies almost entirely within the lower Mekong basin. Its climate, landscape and land use are the major factors shaping the hydrology of the river. The mountainous landscape means that only 16% of the country is farmed under lowland terrace or upland shifting cultivation. With upland shifting agriculture (slash and burn), soils recover within 10 to 20 years but the vegetation does not. Shifting cultivation is common in the uplands of northern Laos and is reported to account for as much as 27% of the total land under rice cultivation. As elsewhere in the basin, forest cover has been steadily reduced during the last three decades by shifting agriculture and permanent agriculture. The cumulative impacts of these activities on the river regime have not been measured. However, the hydrological impacts of land cover changes induced by the Vietnam War were quantified in two sub-catchments of the lower Mekong River basin.
Loss of forest cover in the Thai areas of the lower basin has been the highest of all the lower Mekong countries over the past 60 years. On the Khorat Plateau, which includes the Mun and Chi tributary systems, forest cover was reduced from 42% in 1961 to 13% in 1993. Although this part of northeast Thailand has an annual rainfall of more than 1,000 mm, a high evaporation rate means it is classified as a semi-arid region. Consequently, although the Mun and Chi basins drain 15% of the entire Mekong basin, they only contribute 6% of the average annual flow. Sandy and saline soils are the most common soil types, which makes much of the land unsuitable for wet rice cultivation. In spite of poor fertility, however, agriculture is intensive. Glutinous rice, maize, and cassava are the principal crops. Drought is by far the major hydrological hazard in this region.
As the Mekong enters Cambodia, over 95% of its flows have already joined the river. From here on downstream the terrain is flat and water levels rather than flow volumes determine the movement of water across the landscape. The seasonal cycle of changing water levels at Phnom Penh results in the unique "flow reversal" of water into and out of the Great Lake via the Tonle Sap River. Phnom Penh also marks the beginning of the delta system of the Mekong River. Here the mainstream begins to break up into an increasing number of branches.
In Cambodia, wet rice is the main crop and is grown on the flood plains of the Tonle Sap, Mekong, and Bassac (the Mekong delta distributary known as the Hậu in Vietnam) Rivers. More than half of Cambodia remains covered with mixed evergreen and deciduous broadleaf forest, but forest cover has decreased from 73% in 1973 to 63% in 1993. Here, the river landscape is flat. Small changes in water level determine the direction of water movement, including the large-scale reversal of flow into and out of the Tonle Sap basin from the Mekong River.
The Mekong delta in Vietnam is farmed intensively and has little natural vegetation left. Forest cover is less than 10%. In the Central Highlands of Vietnam, forest cover was reduced from over 95% in the 1950s to around 50% in the mid-1990s. Agricultural expansion and population pressure are the major reasons for land use and landscape change. Both drought and flood are common hazards in the Delta, which many people believe is the most sensitive to upstream hydrological change.
Water flow along its course
Table 1: Country share of Mekong River Basin (MRB) and water flows
By taking into account hydrological regimes, physiography land use, and existing, planned and potential resource developments, the Mekong is divided into six distinct reaches:
Reach 1: Lancang Jiang or Upper Mekong River in China. In this part of the river, the major source of water flowing into the river comes from melting snow on the Tibetan plateau. This volume of water is sometimes called the "Yunnan component" and plays an important role in the low-flow hydrology of the lower mainstream. Even as far downstream as Kratie, the Yunnan component makes up almost 30% of the average dry season flow. A major concern is that the ongoing and planned expansion of dams and reservoirs on the Mekong mainstream in Yunnan could have a significant effect on the low-flow regime of the lower Mekong basin system.
Reach 2: Chiang Saen to Vientiane and Nong Khai. This reach is almost entirely mountainous and covered with natural forest although there has been widespread slash and burn agriculture. Although this reach cannot be termed "unspoiled", the hydrological response is perhaps the most natural and undisturbed of all the lower basin. Many hydrological aspects of the lower basin start to change rapidly at the downstream boundary of this reach.
On 19 July 2019 this reach of the river dropped to its lowest level in a century. Officials are particularly concerned as July is in the wet season, when mainstream flows are abundant historically. Locals are blaming low water on the newly constructed Xayaburi Dam, as it enters its test phase prior to the start of commercial operation in October 2019.
Reach 3: Vientiane and Nong Khai to Pakse. The boundary between Reach 2 and 3 is where the Mekong hydrology starts to change. Reach 2 is dominated in both wet and dry seasons by the Yunnan Component. Reach 3 is increasingly influenced by contributions from the large left bank tributaries in Laos, namely the Nam Ngum, Nam Theun, Nam Hinboun, Se Bang Fai, Se Bang Hieng and Se Done Rivers. The Mun-Chi river system from the right bank in Thailand enters the mainstream within this reach.
Reach 4: Pakse to Kratie. The main hydrological contributions to the mainstream in this reach come from the Se Kong, Se San, and Sre Pok catchments. Together, these rivers make up the largest hydrological sub-component of the lower basin. Over 25% of the mean annual flow volume to the mainstream at Kratie comes from these three river basins. They are the key element in the hydrology of this part of the system, especially to the Tonle Sap flow reversal.
Reach 5: Kratie to Phnom Penh. This reach includes the hydraulic complexities of the Cambodian floodplain, the Tonle Sap and the Great Lake. By this stage, over 95% of the total flow has entered the Mekong system. The focus turns from hydrology and water discharge to the assessment of water level, over- bank storage and flooding and the hydrodynamics that determine the timing, duration and volume of the seasonal flow reversal into and out of the Great Lake.
Reach 6: Phnom Penh to the South China Sea. Here the mainstream divides into a complex and increasingly controlled and artificial system of branches and canals. Key features of flow behaviour are tidal influences and salt water intrusion. Every year, 35–50% of this reach is flooded during the rainy season. The impact of road embankments and similar infrastructure developments on the movement of this flood water is an increasingly important consequence of development.
Table 2 summarises the mean annual flows along the mainstream. The mean annual flow entering the lower Mekong from China is equivalent to a relatively modest 450 mm depth of runoff. Downstream of Vientiane this increases to over 600 mm as the principal left bank tributaries enter the mainstream, mainly the Nam Ngum and Nam Theun. The flow level falls again, even with the right bank entry of the Mun-Chi system from Thailand. Although the Mun–Chi basin drains 20% of the lower system, average annual runoff is only 250 mm. Runoff in the mainstream increases again with the entry from the left bank of the Se Kong from southern Laos and Se San and Sre Pok from Vietnam and Cambodia.
Table 2: Lower Mekong Mainstream annual flow (1960 to 2004) at selected sites.
Flows at Chiang Saen entering the lower basin from Yunnan make up about 15% of the wet season flow at Kratie. This rises to 40% during the dry season, even this far downstream. During the wet season, the proportion of average flow coming from Yunnan rapidly decreases downstream of Chiang Saen, from 70% to less than 20% at Kratie. The dry season contribution from Yunnan is much more significant. The major portion of the balance comes from Laos, which points to a major distinction in the low-flow hydrology of the river. One fraction comes from melting snow in China and Tibet and the rest from over-season catchment storage in the lower basin. This has implications for the occurrence of drought conditions. For example, if runoff from melting snow in any given year is very low, then flows upstream of Vientiane-Nong Khai would be lower.
In a large river system like the Mekong, seasonal flows can be quite variable from year to year. Although the pattern of the annual hydrograph is fairly predictable, its magnitude is not. The average monthly flows along the mainstream are listed in Table 3, providing an indication of their range and variability from year to year. At Pakse, for example, flood season flows during August would exceed 20,000 cubic metres per second nine years out of ten, but exceed 34,000 m3/s only one year in ten.
Table 3: Mekong Mainstream monthly discharge, 1960–2004 (m3/s).
There is little evidence from the last 45 years of data of any systematic changes in the hydrological regime of the Mekong.
Geology
The internal drainage patterns of the Mekong are unusual when compared to those of other large rivers. Most large river systems that drain the interiors of continents, such as the Amazon, Congo, and Mississippi, have relatively simple dendritic tributary networks that resemble a branching tree.
Typically, such patterns develop in basins with gentle slopes where the underlying geological structure is fairly homogenous and stable, exerting little or no control on river morphology. In marked contrast, the tributary networks of the Salween, Yangtze, and particularly the Mekong, are complex with different sub-basins often exhibiting different, and distinct, drainage patterns. These complex drainage systems have developed in a setting where the underlying geological structure is heterogeneous and active, and is the major factor controlling the course of rivers and the landscapes they carve out.
The elevation of the Tibetan Plateau during the Tertiary period was an important factor in the genesis of the south-west monsoon, which is the dominant climatic control influencing the hydrology of the Mekong Basin. Understanding the nature and timing of the elevation of Tibet (and the Central Highlands of Vietnam) therefore helps explain the provenance of sediment reaching the delta and the Tonle Sap Great Lake today. Studies of the provenance of sediments in the Mekong delta reveal a major switch in the source of sediments about eight million years ago (Ma). From 36 to 8 Ma the bulk (76%) of the sediments deposited in the delta came from erosion of the bedrock in the Three Rivers Area. From 8 Ma to the present, however, the contribution from the Three Rivers Area fell to 40%, while that from the Central Highlands rose from 11 to 51%. One of the most striking conclusions of provenance studies is the small contribution of sediment from the other parts of the Mekong basin, notably the Khorat Plateau, the uplands of northern Laos and northern Thailand, and the mountain ranges south of the Three Rivers area.
The last glacial period came to an abrupt end about 19,000 years ago (19 ka) when sea levels rose rapidly, reaching a maximum of about 4.5 m above present levels in the early Holocene about 8 ka. At this time the shoreline of the South China Sea almost reached Phnom Penh and cores recovered from near Angkor Borei contained sediments deposited under the influence of tides, and salt marsh and mangrove swamp deposits. Sediments deposited in the Tonle Sap Great Lake about this time (7.9–7.3 ka) also show indications of marine influence, suggesting a connection to the South China Sea. Although the hydraulic relationships between the Mekong and the Tonle Sap Great Lake systems during the Holocene are not well understood, it is clear that between 9,000 and 7,500 years ago the confluence of the Tonle Sap and the Mekong was in proximity to the South China Sea.
The present river morphology of the Mekong Delta developed over the last 6,000 years. During this period, the delta advanced 200 km over the continental shelf of the South China Sea, covering an area of more than 62,500 km2. From 5.3 to 3.5 ka the delta advanced across a broad embayment formed between higher ground near the Cambodian border and uplands north of Ho Chi Minh City. During this phase of its development the delta was sheltered from the wave action of long-shore currents and was constructed largely through fluvial and tidal processes. At this time the delta was advancing at a rate of 17–18 m per year. After 3.5 ka, however, the delta had built out beyond the embayment and became subject to wave action and marine currents. These deflected deposition south-eastwards in the direction of the Cà Mau Peninsula, which is one of the most recent features of the delta.
For much of its length the Mekong flows through bedrock channels, i.e., channels that are confined or constrained by bedrock or old alluvium in the bed and riverbanks. Geomorphologic features normally associated with the alluvial stretches of mature rivers, such as meanders, oxbow lakes, cut-offs, and extensive floodplains are restricted to a short stretch of the mainstream around Vientiane and downstream of Kratie where the river develops alluvial channels that are free of control exerted by the underlying bedrock.
The Mekong basin is not normally considered a seismically active area as much of the basin is underlain by the relatively stable continental block. Nonetheless, the parts of the basin in northern Laos, northern Thailand, Myanmar and China do experience frequent earthquakes and tremors. The magnitude of these earthquakes rarely exceeds 6.5 on the Richter magnitude scale and is unlikely to cause material damage.
History
The difficulty of navigating the river has meant that it has divided, rather than united, the people who live near it. The earliest known settlements date to 210 BCE, with Ban Chiang being an excellent example of early Iron Age culture. The earliest recorded civilization was the 1st century Indianised-Khmer culture of Funan, in the Mekong delta. Excavations at Oc Eo, near modern An Giang, have found coins from as far away as the Roman Empire. This was succeeded by the Khmer culture Chenla state around the 5th century. The Khmer empire of Angkor was the last great Indianized state in the region. From around the time of the fall of the Khmer empire, the Mekong was the front line between the emergent states of Siam and Tonkin (North Vietnam), with Laos and Cambodia, then on the coast, torn between their influence.
The first European to encounter the Mekong was the Portuguese António de Faria in 1540. A European map of 1563 depicts the river, although even by then little was known of the river upstream of the delta. European interest was sporadic: the Spanish and Portuguese mounted some missionary and trade expeditions, while the Dutch Gerrit van Wuysthoff led an expedition up the river as far as Vientiane in 1641–42.
The French invaded the region in the mid-19th century, capturing Saigon in 1861, and establishing a protectorate over Cambodia in 1863.
The first systematic European exploration began with the French Mekong Expedition led by Ernest Doudard de Lagrée and Francis Garnier, which ascended the river from its mouth to Yunnan between 1866 and 1868. Their chief finding was that the Mekong had too many falls and rapids to ever be useful for navigation. The river's source was found by Pyotr Kuzmich Kozlov in 1900.
From 1893, the French extended their control of the river into Laos, establishing French Indochina by the first decade of the 20th century. This lasted until the First and Second Indochina Wars expelled French from its former colony and defeated US-supported governments.
During the wars in Indochina in the 1970s, a significant quantity of explosives (sometimes, entire barges loaded with military ordnance) sank in the Cambodian section of the Mekong (as well as in the country's other waterways). Besides being a danger for fishermen, unexploded ordnance also creates problems for bridge and irrigation systems construction. As of 2013, Cambodian volunteers are being trained, with the support of the Office of Weapons Removal and Abatement within the US State Department Bureau of Political-Military Affairs, to conduct underwater explosive removal.
In 1995, Laos, Thailand, Cambodia, and Vietnam established the Mekong River Commission (MRC) to manage and coordinate the use and care of the Mekong. In 1996 China and Myanmar became "dialogue partners" of the MRC and the six countries now work together in a cooperative framework. In 2000, the governments of China, Laos, Thailand and Myanmar signed a Agreement on Commercial Navigation on Lancang-Mekong River among the Governments of the People’s Republic of China, the Lao People’s Democratic Republic, the Union of Myanmar and the Kingdom of Thailand which is the mechanism for cooperation with regard to riverine trade on the upper stretches of the Mekong.
Natural history
The Mekong basin is one of the richest areas of biodiversity in the world. Only the Amazon boasts a higher level of bio-diversity. Biota estimates for the Greater Mekong Subregion (GMS) include 20,000 plant species, 430 mammals, 1,200 birds, 800 reptiles and amphibians, and an estimated 850 freshwater fish species (excluding euryhaline species mainly found in salt or brackish water, as well as introduced species). The most species rich orders among the freshwater fish in the river basin are cypriniforms (377 species) and catfish (92 species).
New species are regularly described from the Mekong. In 2009, 145 species previously unknown to science were described from the region, including 29 fish species, two bird species, 10 reptiles, five mammals, 96 plants, and six amphibians. Between 1997 and 2015, an average of two new species per week were discovered in the region. The Mekong Region contains 16 WWF Global 200 ecoregions, the greatest concentration of ecoregions in mainland Asia.
No other river is home to so many species of very large fish. The biggest include three species of Probarbus barbs, which can grow up to and weigh , the giant freshwater stingray (Himantura polylepis, syn. H. chaophraya), which can reach at least in length and in width, the giant pangasius (Pangasius sanitwongsei), giant barb (Catlocarpio siamensis) and the endemic Mekong giant catfish (Pangasianodon gigas). The last three can grow up to about in length and weigh . All of these have declined drastically because of dams, flood control, and overfishing.
One species of freshwater dolphin, the Irrawaddy dolphin (Orcaella brevirostris), was once common in the whole of the lower Mekong but is now very rare, with only 85 individuals remaining.
Among other wetland mammals that have been living in and around the river are the smooth-coated otter (Lutra perspicillata) and fishing cat (Prionailurus viverrinus).
The endangered Siamese crocodile (Crocodylus siamensis) occurs in small isolated pockets within the northern Cambodian and Laotian portions of the Mekong River. The saltwater crocodile (Crocodylus porosus) once ranged from the Mekong Delta up the river into Tonle Sap and beyond but is now extinct in the river, along with being extinct in all of Vietnam and possibly even Cambodia.
Protected areas
The headwaters of the Mekong in Zadoi County, Qinghai, China, are protected in Sanjiangyuan National Nature Reserve. The name Sanjiangyuan means "the sources of the Three Rivers". The reserve also includes the headwaters of the Yellow River and the Yangtze.
The section of the river flowing through deep gorges in Yunnan Province is part of the Three Parallel Rivers of Yunnan Protected Areas, a UNESCO World Heritage Site.
The Tonle Sap Biosphere Reserve in Cambodia contains the largest lake in Southeast Asia. It is a UNESCO Biosphere reserve.
Natural phenomena
The low tide level of the river in Cambodia is lower than the high tide level out at sea, and the flow of the Mekong inverts with the tides throughout its stretch in Vietnam and up to Phnom Penh. The very flat Mekong delta area in Vietnam is thus prone to flooding, especially in the provinces of An Giang and Dong Thap (Đồng Tháp), near the Cambodian border.
Fisheries
Aquatic biodiversity in the Mekong River system is the second highest in the world after the Amazon.
The Mekong boasts the most concentrated biodiversity per hectare of any river.
The commercially valuable fish species in the Mekong are generally divided between "black fish", which inhabit low oxygen, slow moving, shallow waters, and "white fish", which inhabit well oxygenated, fast moving, deeper waters. People living within the Mekong River system generate many other sources of food and income from what are often termed "other aquatic animals" (OAAs) such as freshwater crabs, shrimp, snakes, turtles, and frogs.
OAAs account for about 20% of the total Mekong catch. When fisheries are discussed, catches are typically divided between the wild capture fishery (i.e., fish and other aquatic animals caught in their natural habitat), and aquaculture (fish reared under controlled conditions). Wild capture fisheries play the most important role in supporting livelihoods. Wild capture fisheries are largely open access fisheries, which poor rural people can access for food and income.
Broadly, there are three types of fish habitats in the Mekong: i) the river, including all the main tributaries, rivers in the major flood zone, and the Tonle Sap, which altogether yield about 30% of wild catch landings; ii) rain-fed wetlands outside the river-floodplain zone, including mainly rice paddies in formerly forested areas and usually inundated to about 50 cm, yielding about 66% of wild catch landings; and iii) large water bodies outside the flood zone, including canals and reservoirs yielding about 4% of wild catch landings.
The Mekong Basin has one of the world's largest and most productive inland fisheries. An estimated two million tonnes of fish are landed a year, in addition to almost 500,000 tonnes of other aquatic animals. Aquaculture yields about two million tonnes of fish a year. Hence, the lower Mekong basin yields about 4.5 million tonnes of fish and aquatic products annually. The total economic value of the fishery is between US$3.9 and US$7 billion a year. Wild capture fisheries alone have been valued at US$2 billion a year. This value increases considerably when the multiplier effect is included, but estimates vary widely.
An estimated 2.56 million tonnes of inland fish and other aquatic animals are consumed in the lower Mekong every year. Aquatic resources make up between 47 and 80% of animal protein in rural diets for people who live in the Lower Mekong Basin. Fish are the cheapest source of animal protein in the region and any decline in the fishery is likely to significantly impact nutrition, especially among the poor. Fish are the staple of the diet in Laos and Cambodia, with around 80% of the Cambodian population's annual protein intake coming from fish caught in the Mekong River system, with no alternative source to replace it. An MRC report claims that dam projects on the Mekong River will reduce aquatic life by 40% by 2020, and predicted that 80% of fish will be depleted by 2040. Thailand will be impacted, as its fish stocks in the Mekong will decline by 55%, Laos will be reduced by 50%, Cambodia by 35%, and Vietnam by 30%.
It is estimated that 40 million rural people, more than two-thirds of the rural population in the lower Mekong basin, are engaged in the wild capture fishery. Fisheries contribute significantly to a diversified livelihood strategy for many people, particularly the poor, who are highly dependent on the river and its resources for their livelihoods. They provide a principal form of income for numerous people and act as a safety net and coping strategy in times of poor agricultural harvests or other difficulties. In Laos alone, 71% of rural households (2.9 million people) rely on fisheries for either subsistence or additional cash income. Around the Tonle Sap Lake in Cambodia, more than 1.2 million people live in fishing communes and depend almost entirely on fishing for their livelihoods.
Dams
The Mekong is already heavily dammed, with many more dams planned and under construction, mostly for generating hydropower. China built ten or eleven cascade dams on the Mekong mainstream between 1995 and mid-2019, leaving Vietnam, Laos, Cambodia and Thailand without same amount of water as before investment. China, Laos and Cambodia are planning and/or building more. The Mekong is the fastest growing large river basin in the world in terms of hydropower construction. The Lao Government is aiming to lift the nation out of poverty by making it "the battery of Asia".
Critics fear that China's ability to control the Mekong's flow gives it leverage over downstream nations who rely on China's goodwill. In a worst-case scenario, China could well make demands on thirsty downstream countries who would not be able to refuse. "China could, in short, use its dams to 'weaponize water'".
The Jinghong Dam, the nearest Chinese dam upstream of the Thai border, has caused huge fluctuations in river levels, affecting people's livelihoods downstream by disrupting the river's natural cycle. It is impacting the ecosystem, disturbing the migratory patterns of fish as well as riverbank plants and local agriculture downstream.
Navigation
For thousands of years the Mekong River has been an important conduit for people and goods between the many towns on its banks. Traditional forms of trade in small boats linking communities continue today, however the river is also becoming an important link in international trade routes, connecting the six Mekong countries to each other, and also to the rest of the world. The Mekong is still a wild river and navigation conditions vary greatly along its length. Broadly, navigation of the river is divided between upper and lower Mekong, with the "upper" part of the river defined as the stretch north of the Khone Falls in southern Laos and the "lower" part as the stretch below these falls.
Narrower and more turbulent sections of water in the upstream parts of the Mekong River, coupled with large annual water level variations continue to present a challenge to navigation. The seasonal variations in water level directly affect trade in this section of the river. Volumes of trade being shipped decrease by more than 50%, primarily due to the reduced draughts available during the low water season (June–January). Despite these difficulties, the Mekong River is already an important link in the transit chain between Kunming and Bangkok with about 300,000 tonnes of goods shipped via this route each year. The volume of this trade is expected to increase by 8–11% per year. Port infrastructure is being expanded to accommodate the expected growth in traffic, with new facilities planned for Chiang Saen port.
In Laos, 50 and 100 DWT vessels are operated for regional trade. Cargos carried are timber, agricultural products, and construction materials. Thailand imports a wide variety of products from China, including vegetables, fruit, agricultural products, and fertilisers. The main exports from Thailand are dried longan, fish oil, rubber products, and consumables. Nearly all the ships carrying cargo to and from Chiang Saen Port are 300 DWT Chinese flag vessels.
Waterborne trade in the lower Mekong countries of Vietnam and Cambodia has grown significantly, with trends in container traffic at Phnom Penh port and general cargo through Can Tho port both showing steady increases until 2009 when a decrease in cargo volumes can be attributed to the global financial crisis and a subsequent decline in demand for the export of garments to the US. In 2009, Mekong trade received a significant boost with the opening of a new deep-water port at Cai Mep in Vietnam. This new port has generated a renewed focus on the Mekong River as a trade route. The Cai Mep container terminals can accommodate vessels with a draught of 15.2 m, equivalent to the largest container ships in the world. These mother vessels sail directly to Europe or the United States, which means that goods can be shipped internationally to and from Phnom Penh with only a single transshipment at Cai Mep.
As an international river, a number of agreements exist between the countries that share the Mekong to enable trade and passage between them. The most important of these, which address the full length of the river, are:
Agreement between China and Lao PDR on Freight and Passenger Transport along the Lancang–Mekong River, adopted in November 1994.
Agreement on the Cooperation for the Sustainable Development of the Mekong River Basin, Article 9, Freedom of Navigation, 5 April 1995, Chiang Rai.
Hanoi Agreement between Cambodia and Viet Nam on Waterway Transportation, 13 December 1998.
Agreement between and among the Governments of the Laos, Thailand, and Vietnam for Facilitation of Cross border Transport of Goods and People, (amended at Yangon, Myanmar), signed in Vientiane, 26 November 1999.
Agreement on Commercial Navigation on Lancang–Mekong River among the governments of China, Laos, Myanmar and Thailand, adopted at Tachileik, 20 April 2000.
Phnom Penh Agreement between Cambodia and Vietnam on the Transit of Goods, 7 September 2000.
New Agreement on Waterway Transportation between Vietnam and Cambodia, signed in Phnom Penh, 17 December 2009.
In December 2016, the Thai cabinet of Prime Minister Prayut Chan-o-cha agreed "in principle" to a plan to dredge stretches of the Mekong and demolish rocky outcrops that are hindrances to easy navigation. The international Lancang-Mekong River navigation improvement plan for 2015–2025, conceived by China, Myanmar, Lao, and Thailand, aims to make the river more navigable for 500-tonne cargo vessels sailing the river from Yunnan to Luang Prabang, a distance of 890 kilometres. China has been the driving force behind the demolition plan as it aims to expand trade in the area. The plan is split into two phases. The first phase, from 2015 to 2020, involves a survey, a design, and an assessment of the environmental and social impacts of the project. These have to be approved by the four countries involved: China, Laos, Myanmar, Thailand. The second phase (2020–2025) involves navigational improvements from Simao in China to 243 border posts in China and Myanmar, a distance of 259 km. Local groups have countered that native inhabitants already operate their boats year-round and that the plan to blast the rapids is not about making life better for local people, but about enabling year-round traffic of large Chinese commercial boats.
On 4 February 2020, the Thai Cabinet voted to stop the project to blast and dredge 97 km of the river bed after Beijing failed to stump up the money for further surveys of the affected area.
Bridges
Construction of Myanmar–Laos Friendship Bridge started on 19 February 2013. The bridge will be long and have an wide two-lane motorway.
The Thai–Lao Friendship Bridge (, ) connects Nong Khai city with Vientiane in Laos. The bridge opened on 8 April 1994. It has two lanes with a single railway line in the middle. On 20 March 2004, the Thai and Lao governments agreed to extend the railway to Tha Nalaeng in Laos. This extension has since been completed.
The Second Thai–Lao Friendship Bridge connects Mukdahan to Savannakhet. The two-lane, , bridge opened to the public on 9 January 2007.
The Third Thai–Lao Friendship Bridge opened for traffic on 11 November 2011, connecting Nakhon Phanom Province (Thailand) and Thakhek (Laos), as part of Asian Highway 3. The Chinese and Thai governments agreed to build the bridge and share the estimated US$33 million cost.
The Fourth Thai–Lao Friendship Bridge opened to traffic on 11 December 2013. It links Chiang Rai Province, Thailand with Ban Houayxay, Laos.
There is one bridge over the Mekong entirely within Laos. Unlike the Friendship Bridges, it is not a border crossing. It is at Pakse in Champasak Province. It is long, and was completed in 2000. ).
The Kizuna Bridge is in Cambodia, in the city of Kampong Cham, on the road linking Phnom Penh with the remote provinces of Ratanakiri and Mondolkiri, and Laos. The bridge opened for traffic on 11 December 2001.
The Prek Tamak Bridge, north of Phnom Penh opened in 2010.
Phnom Penh itself has no bridge under construction yet, although two new bridges have recently opened on the Tonle Sap, and the main bridge on the highway to Ho Chi Minh was duplicated in 2010.
Another new bridge was built at Neak Leung on the Phnom Penh to Ho Chi Minh Highway 1 with Japanese government assistance, and opened in 2015.
In Vietnam, since the year 2000 Mỹ Thuận Bridge crosses the first channel—the left, main branch of the Mekong, the Sông Tiền or Tiền Giang—near Vĩnh Long and since 2008 Rạch Miễu Bridge crosses it near Mỹ Tho, between the provinces of Tiền Giang and Bến Tre.
Cần Thơ Bridge crosses the second channel—the right, main distributary of the Mekong, the Bassac (Song Hau). Inaugurated in 2010, it is the longest main span cable-stayed bridge in Southeast Asia.
Environmental issues
Drought linked to a changing climate and dozens of hydroelectric dams are damaging the Mekong ecosystem. When drought ends and the inevitable floods begin, the effects of Mekong dams on flood pulse dynamics over the entire Lower Mekong are poorly understood.
Sewage treatment is rudimentary in towns and urban areas throughout much of the Mekong's length, such as Vientiane in Laos. Water pollution impacts the river's ecological integrity as a result.
Much of the 8.3 billion tonnes of plastic present on earth makes its way to the oceans. Ninety percent of plastic in the oceans is flushed there by just 10 rivers. The Mekong is one of them.
A growing number of academics, NGOs, and scientists urge the international community and the Mekong River Commission to put to rest the myth of sustainable hydro power. They urge an immediate moratorium on new construction and a shift to solar and other forms of renewable energy, already highly competitive and faster to install.
See also
Greater Mekong Subregion
Greater Mekong Sub-region Academic and Research Network
GMS Environment Operations Center
Indochina
Sekong River
Mekong River Basin Hydropower
Mekong River massacre 2011 killings on Mekong river
Stung Sen River
Mekong Expedition of 1866-1868
Fair river sharing
References
Further reading
Kuenzer, C., I. Campbell, M. Roch, P. Leinenkugel, V. Q. Tuan and S. Dech (2012): Understanding the impact of hydropower developments in the context of upstream–downstream relations in the Mekong river basin. In: Sustainability Science 8 (4), pp. 565–584. DOI: 10.1007/s11625-012-0195-z.
Kuenzer, C., H. Guo, J. Huth, P. Leinenkugel, X. Li and S. Dech (2013): Flood Mapping and Flood Dynamics of the Mekong Delta. ENVISAT-ASAR-WSM Based Time-Series Analyses. In: Remote Sensing 5, pp. 687–715. DOI: 10.3390/rs5020687.
Leinenkugel, P., C. Kuenzer, N. Oppelt and S. Dech (2013): Characterisation of land surface phenology and land cover based on moderate resolution satellite data in cloud prone areas – A novel product for the Mekong Basin. In: Remote Sensing of Environment 136, pp. 180–198. DOI: 10.1016/j.rse.2013.05.004.
Moder, F., C. Kuenzer, Z. Xu, P. Leinenkugel and Q. Bui Van (2012): IWRM for the Mekong Basin. In: Renaud, F. G. and C. Kuenzer (eds.): The Mekong Delta System. Interdisciplinary Analyses of a River Delta. Dordrecht: Springer, pp. 133–166.
Renaud, F. G. und C. Kuenzer (2012): The Mekong Delta System. Interdisciplinary Analyses of a River Delta (= Springer Environmental Science and Engineering). Dordrecht: Springer. .
Kuenzer, C. and F. G. Renaud (2012): Climate Change and Environmental Change in River Deltas Globally. In: Renaud, F. G. and C. Kuenzer (eds.): The Mekong Delta System. Interdisciplinary Analyses of a River Delta. Dordrecht: Springer, pp. 7–48.
Kuenzer, C. (2014): Remote Sensing the Mekong. In: International Journal of Remote Sensing 35 (8), pp. 2747–2751. DOI: 10.1080/01431161.2014.890377.
Kuenzer, C., P. Leinenkugel, M. Vollmuth and S. Dech (2014): Comparing global land-cover products – implications for geoscience applications: an investigation for the trans-boundary Mekong Basin. In: International Journal of Remote Sensing 35 (8), pp. 2752–2779. DOI: 10.1080/01431161.2014.890305.
External links
The WISDOM Project, a Water related Information System for the Mekong Delta
Mekong River Commission
CGIAR Research Program on Water, Land and Ecosystems - Greater Mekong
Mekong Watch
Countries of the Mekong River: The Greater Mekong Subregion Asian Development Bank
Rivers Network : Mekong river blog
Ecoregions of China
Freshwater ecoregions
International rivers of Asia
Rivers of Tibet
Rivers of Yunnan
Rivers of Cambodia
Rivers of Laos
Rivers of Myanmar
Rivers of Thailand
Rivers of Vietnam
Tonlé Sap
Isan
Laos–Thailand border
Laos–Myanmar border
Cambodia–Laos border
Border rivers
Ramsar sites in Cambodia
Ecoregions of Laos
Ecoregions of Myanmar
Ecoregions of Thailand
Ecoregions of Cambodia
Ecoregions of Vietnam |
null | null | Vasectomy | eng_Latn | Vasectomy is a surgical procedure for male sterilization or permanent contraception. During the procedure, the male vasa deferentia are cut and tied or sealed so as to prevent sperm from entering into the urethra and thereby prevent fertilization of a female through sexual intercourse. Vasectomies are usually performed in a physician's office, medical clinic, or, when performed on an animal, in a veterinary clinic. Hospitalization is not normally required as the procedure is not complicated, the incisions are small, and the necessary equipment routine.
There are several methods by which a surgeon might complete a vasectomy procedure, all of which occlude (i.e., "seal") at least one side of each vas deferens. To help reduce anxiety and increase patient comfort, men who have an aversion to needles may consider a "no-needle" application of anesthesia while the 'no-scalpel' or 'open-ended' techniques help to accelerate recovery times and increase the chance of healthy recovery.
Due to the simplicity of the surgery, a vasectomy usually takes less than 30 minutes to complete. After a short recovery at the doctor's office (usually less than an hour), the patient is sent home to rest. Because the procedure is minimally invasive, many vasectomy patients find that they can resume their typical sexual behavior within a week, and do so with little or no discomfort.
Because the procedure is considered a permanent method of contraception and is not easily reversed, men are usually counseled and advised to consider how the long-term outcome of a vasectomy might affect them both emotionally and physically. The procedure is not often encouraged for young single childless men as their chances for biological parenthood are thereby permanently reduced, sometimes completely.
Medical uses
A vasectomy is done to prevent fertility in males. It ensures that in most cases the person will be sterile after confirmation of success following surgery. The procedure is regarded as permanent because vasectomy reversal is costly and often does not restore the male's sperm count or sperm motility to prevasectomy levels. Men with vasectomies have a very small (nearly zero) chance of successfully impregnating a woman, but a vasectomy has no effect on rates of sexually transmitted infections.
After vasectomy, the testes remain in the scrotum where Leydig cells continue to produce testosterone and other male hormones that continue to be secreted into the bloodstream. Some studies have found that sexual desire after vasectomy may be somewhat diminished.
When the vasectomy is complete, sperm cannot exit the body through the penis. Sperm is still produced by the testicles but is broken down and absorbed by the body. Much fluid content is absorbed by membranes in the epididymis, and much solid content is broken down by the responding macrophages and reabsorbed via the bloodstream. Sperm is matured in the epididymis for about a month before leaving the testicles. After vasectomy, the membranes must increase in size to absorb and store more fluid; this triggering of the immune system causes more macrophages to be recruited to break down and reabsorb more solid content. Within one year after vasectomy, sixty to seventy percent of vasectomized men develop antisperm antibodies. In some cases, vasitis nodosa, a benign proliferation of the ductular epithelium, can also result. The accumulation of sperm increases pressure in the vas deferens and epididymis. The entry of the sperm into the scrotum can cause sperm granulomas to be formed by the body to contain and absorb the sperm which the body will treat as a foreign biological substance (much like a virus or bacterium).
Efficacy
Vasectomy is the most effective permanent form of contraception available to men. (Removing the entire vas deferens would very likely be more effective, but it is not something that is regularly done.) In nearly every way that vasectomy can be compared to tubal ligation it has a more positive outlook. Vasectomy is more cost effective, less invasive, has techniques that are emerging that may facilitate easier reversal, and has a much lower risk of postoperative complications.
Early failure rates, i.e. pregnancy within a few months after vasectomy, typically result from unprotected sexual intercourse too soon after the procedure while some sperm continue to pass through the vasa deferentia. Most physicians and surgeons who perform vasectomies recommend one (sometimes two) postprocedural semen specimens to verify a successful vasectomy; however, many men fail to return for verification tests citing inconvenience, embarrassment, forgetfulness, or certainty of sterility. In January 2008, the FDA cleared a home test called SpermCheck Vasectomy that allows patients to perform postvasectomy confirmation tests themselves; however, compliance for postvasectomy semen analysis in general remains low.
Late failure, i.e. pregnancy following spontaneous recanalization of the vasa deferentia, has also been documented. This occurs because the epithelium of the vas deferens (similar to the epithelium of some other human body parts) is capable of regenerating and creating a new tube if the vas deferens is damaged and/or severed. Even when as much as five centimeters (or two inches) of the vas deferens is removed, the vas deferens can still grow back together and become reattached—thus allowing sperm to once again pass and flow through the vas deferens, restoring one's fertility.
The Royal College of Obstetricians and Gynaecologists states there is a generally agreed-upon rate of late failure of about one in 2000 vasectomies— better than tubal ligations for which the failure rate is one in every 200 to 300 cases. A 2005 review including both early and late failures described a total of 183 recanalizations from 43,642 vasectomies (0.4%), and sixty pregnancies after 92,184 vasectomies (0.07%).
Complications
Short-term possible complications include infection, bruising and bleeding into the scrotum resulting in a collection of blood known as a hematoma. A study in 2012 demonstrated an infection rate of 2.5% postvasectomy. The stitches on the small incisions required are prone to irritation, though this can be minimized by covering them with gauze or small adhesive bandages. The primary long-term complications are chronic pain conditions or syndromes that can affect any of the scrotal, pelvic or lower-abdominal regions, collectively known as post-vasectomy pain syndrome. Though vasectomy results in increases in circulating immune complexes, these increases are transient. Data based on animal and human studies indicate these changes do not result in increased incidence of atherosclerosis. The risk of testicular cancer is not affected by vasectomy.
In 2014 the AUA reaffirmed that vasectomy is not a risk factor for prostate cancer and that it is not necessary for physicians to routinely discuss prostate cancer in their preoperative counseling of vasectomy patients. There remains ongoing debate regarding whether vasectomy is associated with prostate cancer. A 2017 meta-analysis found no statistically significant increase in risk. A 2019 study of 2.1 million Danish men found that vasectomy increased their incidence of prostate cancer by 15%. A 2020 meta-analysis found that vasectomy increased the incidence by 9%. Other recent studies agree on the 15% increase in risk of developing prostate cancer, but found that men who get a vasectomy are not more likely to die from prostate cancer than men without a vasectomy.
Postvasectomy pain
Post-vasectomy pain syndrome (PVPS) is a chronic and sometimes debilitating condition that may develop immediately or several years after vasectomy. The most robust study of post-vasectomy pain, according to the American Urology Association's Vasectomy Guidelines 2012 (amended 2015) found a rate of 14.7% reported new-onset scrotal pain at 7 months after vasectomy with 0.9% describing the pain as "quite severe and noticeably affecting their quality of life". The pain can be constant orchialgia or epididymal pain (epididymitis), or it can be pain that occurs only at particular times such as with sexual intercourse, ejaculation, or physical exertion.
Psychological effects
Approximately 90% are generally reported in reviews as being satisfied with having had a vasectomy, while 7–10% of men regret their decision. For men in relationships, regret was less common when both people in the relationship agreed on the procedure.
Men who are of a younger age at the time of having a vasectomy are significantly more likely to regret and seek a reversal of their vasectomy, with one study showing men for example in their twenties being 12.5 times more likely to undergo a vasectomy reversal later in life (and including some who chose sterilization at a young age). Pre-vasectomy counseling is therefore of particular importance in younger patients.
Dementia
An association between vasectomy and primary progressive aphasia, a rare variety of frontotemporal dementia, was reported. However, it is doubtful that there is a causal relationship. The putative mechanism is a cross-reactivity between brain and sperm, including the shared presence of neural surface antigens. In addition, the cytoskeletal tau protein has been found only to exist outside of the CNS in the manchette of sperm.
Procedure
The traditional incision approach of vasectomy involves numbing of the scrotum with local anesthetic (although some men's physiology may make access to the vas deferens more difficult in which case general anesthesia may be recommended) after which a scalpel is used to make two small incisions, one on each side of the scrotum at a location that allows the surgeon to bring each vas deferens to the surface for excision. The vasa deferentia are cut (sometimes a section may be removed altogether), separated, and then at least one side is sealed by ligating (suturing), cauterizing (electrocauterization), or clamping. There are several variations to this method that may improve healing, effectiveness, and which help mitigate long-term pain such as post-vasectomy pain syndrome (PVPS) or epididymitis, however the data supporting one over another are limited.
Fascial interposition: Recanalization of the vas deferens is a known cause of vasectomy failure(s). Fascial interposition ("FI"), in which a tissue barrier is placed between the cut ends of the vas by suturing, may help to prevent this type of failure, increasing the overall success rate of vasectomy while leaving the testicular end within the confines of the fascia. The fascia is a fibrous protective sheath that surrounds the vas deferens as well as all other body muscle tissue. This method, when combined with intraluminal cautery (where one or both sides of the vas deferens are electrically "burned" closed to prevent recanalization), has been shown to increase the success rate of vasectomy procedures.
No-needle anesthesia: Fear of needles for injection of local anesthesia is well known. In 2005, a method of local anesthesia was introduced for vasectomy which allows the surgeon to apply it painlessly with a special jet-injection tool, as opposed to traditional needle application. The numbing agent is forced/pushed onto and deep enough into the scrotal tissue to allow for a virtually pain-free surgery. Lidocaine applied in this manner achieves anesthesia in less than one minute. Initial surveys show a very high satisfaction rate amongst vasectomy patients. Once the effects of no-needle anesthesia set in, the vasectomy procedure is performed in the routine manner. However, unlike in conventional local anesthesia where needles and syringes are used on one patient only, the applicator is not single use and cannot be properly cleaned leading to concerns regarding infection control.
No-scalpel vasectomy (NSV): Also known as a "key-hole" vasectomy, is a vasectomy in which a sharp hemostat (as opposed to a scalpel) is used to puncture the scrotum. This method has come into widespread use as the resulting smaller "incision" or puncture wound typically limits bleeding and hematomas. Also the smaller wound has less chance of infection, resulting in faster healing times compared to the larger/longer incisions made with a scalpel. The surgical wound created by the no-scalpel method usually does not require stitches. NSV is the most commonly performed type of minimally invasive vasectomy, and both describe the method of vasectomy that leads to access of the vas deferens.
Open-ended vasectomy: In this procedure the testicular end of the vas deferens is not sealed, which allows continued streaming of sperm into the scrotum. This method may avoid testicular pain resulting from increased back-pressure in the epididymis. Studies suggest that this method may reduce long-term complications such as post-vasectomy pain syndrome.
Vas irrigation: Injections of sterile water or euflavine (which kills sperm) are put into the distal portion of the vas at the time of surgery which then brings about a near-immediate sterile ("azoospermatic") condition. The use of euflavine does however, tend to decrease time (or, number of ejaculations) to azoospermia vs. the water irrigation by itself. This additional step in the vasectomy procedure, (and similarly, fascial interposition), has shown positive results but is not as prominently in use, and few surgeons offer it as part of their vasectomy procedure.
Other techniques
The following vasectomy methods have purportedly had a better chance of later reversal but have seen less use by virtue of known higher failure rates (i.e., recanalization). An earlier clip device, the VasClip, is no longer on the market, due to unacceptably high failure rates.
The VasClip method, though considered reversible, has had a higher cost and resulted in lower success rates. Also, because the vasa deferentia are not cut or tied with this method, it could technically be classified as other than a vasectomy. Vasectomy reversal (and the success thereof) was conjectured to be higher as it only required removing the Vas-Clip device. This method achieved limited use, and scant reversal data are available.
Vas occlusion techniques
Injected plugs: There are two types of injected plugs which can be used to block the vasa deferentia. Medical-grade polyurethane (MPU) or medical-grade silicone rubber (MSR) starts as a liquid polymer that is injected into the vas deferens after which the liquid is clamped in place until is solidifies (usually in a few minutes).
Intra-vas device: The vasa deferentia can also be occluded by an intra-vas device (IVD). A small cut is made in the lower abdomen after which a soft silicone or urethane plug is inserted into each vas tube thereby blocking (occluding) sperm. This method allows for the vas to remain intact. IVD technique is done in an out-patient setting with local anesthetic, similar to a traditional vasectomy. IVD reversal can be performed under the same conditions making it much less costly than vasovasostomy which can require general anesthesia and longer surgery time.
Both vas occlusion techniques require the same basic patient setup: local anesthesia, puncturing of the scrotal sac for access of the vas, and then plug or injected plug occlusion. The success of the aforementioned vas occlusion techniques is not clear and data are still limited. Studies have shown, however, that the time to achieve sterility is longer than the more prominent techniques mentioned in the beginning of this article. The satisfaction rate of patients undergoing IVD techniques has a high rate of satisfaction with regard to the surgery experience itself.
Recovery
Sexual intercourse can usually be resumed in about a week (depending on recovery); however, pregnancy is still possible as long as the sperm count is above zero. Another method of contraception must be relied upon until a sperm count is performed either two months after the vasectomy or after ten to twenty ejaculations have occurred.
After a vasectomy, contraceptive precautions must be continued until azoospermia is confirmed. Usually two semen analyses at three and four months are necessary to confirm azoospermia. The British Andrological Society has recommended that a single semen analysis confirming azoospermia after sixteen weeks is sufficient.
Post-vasectomy, testicles will continue to produce sperm cells. As before vasectomy, unused sperm are reabsorbed by the body.
Conceiving after vasectomy
In order to allow the possibility of reproduction via artificial insemination after vasectomy, some men opt for cryopreservation of sperm before sterilization. It is advised that all men having a vasectomy consider freezing some sperm before the procedure. Dr Allan Pacey, senior lecturer in andrology at Sheffield University and secretary of the British Fertility Society, notes that men who he sees for a vasectomy reversal which has not worked express wishing they had known they could have stored sperm. Pacey notes, "The problem is you're asking a man to foresee a future where he might not necessarily be with his current partner—and that may be quite hard to do when she's sitting next to you."
The cost of cryo-preservation (sperm banking) may also be substantially less than alternative vaso-vasectomy procedures, compared to the costs of in-vitro fertilization (IVF) which usually run from $12,000 to $25,000.
Sperm can be aspirated from the testicles or the epididymis, and while there is not enough for successful artificial insemination, there is enough to fertilize an ovum by intracytoplasmic sperm injection. This avoids the problem of antisperm antibodies and may result in a faster pregnancy. IVF may be less costly per cycle than reversal in some health-care systems, but a single IVF cycle is often insufficient for conception. Disadvantages include the need for procedures on the woman, and the standard potential side-effects of IVF for both the mother and the child.
Vasectomy reversal
Although men considering vasectomies should not think of them as reversible, and most men and their partners are satisfied with the operation, life circumstances and outlooks can change, and there is a surgical procedure to reverse vasectomies using vasovasostomy (a form of microsurgery first performed by Earl Owen in 1971). Vasovasostomy is effective at achieving pregnancy in a variable percentage of cases, and total out-of-pocket costs in the United States are often upwards of $10,000. The typical success rate of pregnancy following a vasectomy reversal is around 55% if performed within 10 years, and drops to around 25% if performed after 10 years. After reversal, sperm counts and motility are usually much lower than pre-vasectomy levels. There is evidence that men who have had a vasectomy may produce more abnormal sperm, which would explain why even a mechanically successful reversal does not always restore fertility. The higher rates of aneuploidy and diploidy in the sperm cells of men who have undergone vasectomy reversal may lead to a higher rate of birth defects.
Some reasons that men seek vasectomy reversals include wanting a family with a new partner following a relationship breakdown or divorce, their original partner dying and subsequently going on to repartner and to want children, the unexpected death of a child, or a long-standing couple changing their minds some time later, often prompted by situations such as improved finances or existing children approaching the age of school or leaving home. Patients often comment that they never anticipated the possibility of a relationship breakdown or death, or how that might affect their situation at the time of having their vasectomy. A small number of vasectomy reversals are also performed in attempts to relieve post-vasectomy pain syndrome.
Prevalence
Internationally, vasectomy rates are vastly different. While female sterilisation is the most widely used method worldwide, with 223 million women relying on it, only 28 million women rely on their partner's vasectomy. In the world's 69 least developed countries less than 0.1% of men use vasectomies on average. Of 54 African countries, only ten report measurable vasectomy use and only Swaziland, Botswana, and South Africa exceed 0.1% prevalence.
In North America and Europe vasectomy usage is on the order of 10% with some countries reaching 20%. Despite its high efficacy, in the United States, vasectomy is utilized less than half the rate of the alternative female tubal ligation. According to the research, vasectomy in the US is least utilized among black and Latino populations, the groups that have the highest rates of female sterilization.
New Zealand, in contrast, has higher levels of vasectomy than tubal ligation. 18% of all men, and 25% of all married men, have had a vasectomy. The age cohort with the highest level of vasectomy was 40–49, where 57% of men had taken it up. Canada, the UK, Bhutan and the Netherlands all have similar levels of uptake.
History
The first recorded vasectomy was performed on a dog in 1823. A short time after that, R. Harrison of London performed the first human vasectomy; however, the surgery was done not for sterilization purposes, but to bring about atrophy of the prostate. Soon, however, it was believed to have benefits for eugenics. The first case report of vasectomy in the United States was in 1897, by A.J. Ochsner, a surgeon in Chicago, in a paper titled, "Surgical treatment of habitual criminals." He believed vasectomy to be a simple, effective means for stemming the tide of racial degeneration widely believed to be occurring. In 1902, Harry C. Sharp, the surgeon at the Indiana Reformatory, reported that he had sterilized forty-two inmates in an effort to both reduce criminal behavior in those individuals and prevent the birth of future criminals.
Not all vasectomies were performed with the goal of sterilization. Eugen Steinach (1861–1944), an Austrian physician, believed that a unilateral vasectomy (severing only one of the two vasa deferentia) in older men could restore general vigor and sexual potency, shrink enlarged prostates, and cure various ailments by somehow boosting the hormonal output of the vasectomized testicle. This surgery, which became very popular in the 1920s, was undertaken by many wealthy men, including Sigmund Freud and W. B. Yeats. Since these operations lacked rigorous controlled trials, any rejuvenating effect was likely due to the placebo effect, and with the later development of synthetic injectable hormones, this operation fell out of vogue.
Vasectomy began to be regarded as a method of consensual birth control during the Second World War. The first vasectomy program on a national scale was launched in 1954 in India.
The procedure is seldom performed on dogs, with castration remaining the preferred reproductive control option for canines. It is regularly performed on bulls.
Society and culture
Availability and legality
Vasectomy costs are (or may be) covered in different countries, as a method of both contraception or population control, with some offering it as a part of a national health insurance. The Affordable Care Act of the U.S. does not cover vasectomy. Vasectomy was generally considered illegal in France until 2001, due to provisions in the Napoleonic Code forbidding "self-mutilation". No French law specifically mentioned vasectomy until a 2001 law on contraception and infanticide permitted the procedure. In 2014, the Iranian parliament voted for a bill that would ban vasectomies.
Tourism
Medical tourism, where a patient travels to a less developed location where a procedure is cheaper to save money and combine convalescence with a vacation, is infrequently used for vasectomy due to its low cost, but is more likely to be used for vasectomy reversal. Many hospitals list vasectomy as being available. Medical tourism has been scrutinized by some governments for quality of care and postoperative care issues.
Shooting of Dr. Andrew Rynne
In 1990, chairperson of the Irish Family Planning Association, and the Republic of Ireland's first vasectomy specialist, Andrew Rynne, was shot by a former client, but he survived. The incident is the subject of a short film, The Vasectomy Doctor, by Paul Webster.
See also
Male contraceptive
Sperm granuloma
Testicular sperm extraction
Orchiectomy
List of surgeries by type
World Vasectomy Day
References
External links
MedlinePlus Encyclopedia
How to treat: vasectomy and reversal, Australian Doctor, 2 July 2014
Sterilization (medicine)
Contraception for males
1897 introductions
Male genital surgery
Vasectomy |
null | null | Morning Star | eng_Latn | Morning Star, morning star, or Morningstar may refer to:
Astronomy
Morning star, most commonly used as a name for the planet Venus when it appears in the east before sunrise
See also Venus in culture
Morning star, a name for the star Sirius, which appears in the sky just before sunrise from early July to mid-September
Morning star, a (less common) name for the planet Mercury when it appears in the east before sunrise
Mythology and theology
At-Tariq, a chapter of the Quran
Barnumbirr, a creator-spirit in the Yolngu culture of Australia
Jesus, self-described as "the bright Morning Star" in the Christian Bible
John the Baptist, called a "bright morning star" in Eastern Orthodox Church hymnology
Lucifer, a name based on the Latin name for the Morning Star
Mary, mother of Jesus, called "morning star" in the Litany of Loreto
Morning Star, one of the Zorya (gods in Slavic mythology)
Morningstar or Red Horn, a culture hero in Siouan oral traditions
Morning Star ceremony, an historical ritual ceremony in Pawnee mythology
Morning Star ceremony, a Yolngu (Aboriginal Australian) ceremony associated with Barnumbirr
Phosphorus (morning star), the Morning Star in Greek mythology
Tlāhuizcalpantecuhtli, god of the morning star (Venus) Aztec mythology
People
John Wycliffe, English theologian, sometimes referred to as the "Morning Star of the Reformation"
Morning Star (chief) (1810–1883), Northern Cheyenne leader, also known as Dull Knife
Chip Morningstar, American author, developer, programmer and designer of software systems
Darren Morningstar (born 1969), American basketball player
Erica Morningstar (born 1989), Canadian swimmer
Richard Morningstar (born 1945), ambassador, Special Envoy of the United States Secretary of State for Eurasian Energy
Places
Morning Star, Arkansas, an unincorporated community
Morning Star, Virginia, an unincorporated community
Morning Star, West Virginia, an unincorporated community
Morning Star Lake (Nebraska)
Morningstar, U.S. Virgin Islands
Morningstar Airfield, near Cape Town, Western Cape, South Africa
Morningstar Mill, a heritage site in St. Catharines, Ontario, Canada
Transportation
Ships
HMS Morning Star, six ships of the Royal Navy
Morning Star, original name of the USS Estes, an amphibious force command ship
Morning Star (ship), wrecked in the Torres Strait in 1814
Morning Star, former name of the cruise ship Salamis Glory
Morning Star of Revelation (TS-K182), a British sail-training ketch
Morning Star, the first Star Ferry in Hong Kong, and several later ferries by the same name in the same fleet
De Ochtendster (The Morning Star in Dutch), took part in the Battle of Gibraltar (1607)
In railways
Morning Star (train), passenger train formerly operated by St. Louis Southwestern Railway
GWR Star Class "Morning Star", broad gauge steam locomotive
Businesses and organizations
Businesses
Morningstar Farms, a brand of vegetarian food
Morning Star Series, a brand of incense
Morningstar Air Express, a cargo airline based in Edmonton, Alberta, Canada
Morningstar, Inc., a Chicago-based investment research firm
The Morning Star Company, a California-based agribusiness and food processing company
Organizations
Morningstar Commune, formed in Sonoma County in the mid-1960s
Morning Star Boys' Ranch, Spokane, Washington
Morning Star Trust, U.K.-based, Christian sail training charity
Morning Star International, former name of Every Nation Churches, an organization of non-denominational churches
Arts and entertainment
Literature
Comics
Morning Star (comics), a fictional Marvel Comics villain
Lucifer (DC Comics) (Lucifer Morningstar), a DC character who primarily appears in The Sandman comic books by Neil Gaiman
Morningstar, a superhero from the Bogatyri comic books
Morningstar, a character in the Elementals comic book
Newsmedia
Morning Star (British newspaper), a national daily newspaper in the U.K.
Morning Star (London newspaper), a defunct London newspaper
The Morning Star (British Columbia newspaper), a daily newspaper in Vernon, British Columbia
The Morning Star (New Hampshire newspaper), a defunct New Hampshire newspaper
Morning Star, established 1867 in Wilmington, North Carolina, renamed Star-News in 2003
Magic City Morning Star, an on-line newspaper from Katahdin, Maine
Novels
Marjorie Morningstar (novel), a 1955 novel by Herman Wouk
Morning Star (Haggard novel), a 1910 novel by H Rider Haggard
Morning Star (Raven novel), a 1984 novel by Simon Raven
Morning Star (Brown novel), a 2016 novel by Pierce Brown
The Morning Star (novel), 2020 novel by Karl Ove Knausgård
Morningstar, a 1992 novel by David Gemmell
The Morning Star, a 2001–2003 trilogy of novels by Nick Bantock
Morning Star, a character in the Noble Warriors Trilogy of novels by William Nicholson
Morningstar, a character from the novel Jack of Shadows by Roger Zelazny
Other literature
The Morning Star, a play by Emlyn Williams
The Morning Star, a 1979 book of poems and translations by Kenneth Rexroth
"Luceafărul" (poem), an 1883 narrative poem sometimes translated as "The Morning Star"
Film and television
Marjorie Morningstar (film), a 1958 film based on the novel
Morning Star (film), a 1959 Soviet ballet film
Morning Star (TV series), a short-lived 1965–1966 American soap opera that was broadcast on NBC
Morning Star (talk show), a 2004–2005 Philippine morning talk show that aired on ABS-CBN
Morning Star, a planned film about Djalu Gurruwiwi
Lucifer Morningstar, a character in the TV series Lucifer
Music
Morning Star (opera), by Ricky Ian Gordon
"Morning Star" (carol), an 1836 traditional Moravian carol
The Morningstars, a band fronted by Sierra Leonean musician S. E. Rogie
Albums
Morning Star (Entombed album), 2001
Morning Star (Flunk album), 2004
Morning Star (Hubert Laws album), 1972
Morning Star (Spear of Destiny album), 2003
Morningstar, an album by the American electronic music band Asmodeus X
The Morning Star (Daniel Bachman album), 2018
Criminal Intents/Morning Star, a 2009 EP by the band Dope Stars Inc.
Songs
"Morningstar", by AFI on the album The Art of Drowning
"Morning Star", by Amorphis on the album Tuonela
"Morning Star", by Angra on the album Temple of Shadows
"Morningstar", by Baroness on the album Purple
"Morning Star", by Cass McCombs on the album Big Wheel and Others
"Morning Star", a musical work for wind ensemble composed by David Maslanka
"The Morningstar", by Draconian on the album The Burning Halo
"Morning Star, Durbanville", by Fokofpolisiekar on the album Swanesang
"Morning Star", by Hank Marvin
"Morning Star" (Nat King Cole song), 1958
"Morning Star" (N-Dubz song), 2011
"Morning Star", by Roger Shah featuring the voice of Moya Brennan
"Morning Star", by Stiff Gins
"Morning Star", by Taylor Davis
"Draconian Trilogy: Morning Star", by Therion on the album Vovin
"Morning Star", by Amberian Dawn on the album The Clouds of Northland Thunder
"Morning Star", by King Woman
Other arts
Morning Star, a painting by Alex Janvier
Other uses
Morning Star (cannabis), a strain of Cannabis sativa
Morning star (candlestick pattern), a formation in a financial candlestick chart
Morning star (weapon), a spiked mace
Morning Star flag, the national flag of West Papua
See also
The Evening and the Morning Star (1832–1834), the first newspaper of the Latter Day Saint movement
Evening Star (disambiguation)
Morgenster (disambiguation)
Morgenstern (disambiguation) |
null | null | Victor Emmanuel II of Italy | eng_Latn | Victor Emmanuel II (; full name: Vittorio Emanuele Maria Alberto Eugenio Ferdinando Tommaso di Savoia; 14 March 1820 – 9 January 1878) was King of Sardinia from 1849 until 17 March 1861, when he assumed the title of King of Italy and became the first king of an independent, united Italy since the 6th century, a title he held until his death in 1878. Borrowing from the old Latin title Pater Patriae of the Roman emperors, the Italians gave him the epithet of Father of the Fatherland ().
Born in Turin as the eldest son of Charles Albert, Prince of Carignano, and Maria Theresa of Austria, he fought in the First Italian War of Independence (1848–1849) before being made King of Piedmont-Sardinia following his father's abdication. He appointed Camillo Benso, Count of Cavour, as his Prime Minister, and he consolidated his position by suppressing the republican left. In 1855, he sent an expeditionary corps to side with French and British forces during the Crimean War; the deployment of Italian troops to the Crimea, and the gallantry shown by them in the Battle of the Chernaya (16 August 1855) and in the siege of Sevastopol led the Kingdom of Sardinia to be among the participants at the peace conference at the end of the war, where it could address the issue of the Italian unification to other European powers. This allowed Victor Emmanuel to ally himself with Napoleon III, Emperor of France. France had supported Sardinia in the Second Italian War of Independence, resulting in liberating Lombardy from Austrian rule.
Victor Emmanuel supported the Expedition of the Thousand (1860–1861) led by Giuseppe Garibaldi, which resulted in the rapid fall of the Kingdom of the Two Sicilies in southern Italy. However, Victor Emmanuel halted Garibaldi when he appeared ready to attack Rome, still under the Papal States, as it was under French protection. In 1860, Tuscany, Modena, Parma and Romagna decided to side with Sardinia-Piedmont, and Victor Emmanuel then marched victoriously in the Marche and Umbria after the victorious Battle of Castelfidardo over the Papal forces. This led to his excommunication from the Catholic Church until 1878, just before his death in the same year. He subsequently met Garibaldi at Teano, receiving from him the control of southern Italy and becoming the first King of Italy on 17 March 1861.
In 1866, the Third Italian War of Independence allowed Italy to annex Veneto. In 1870, Victor Emmanuel also took advantage of the Prussian victory over France in the Franco-Prussian War to conquer the Papal States after the French withdrew. He entered Rome on 20 September 1870 and set up the new capital there on 2 July 1871. He died in Rome in 1878, and was buried in the Pantheon.
The Italian national Victor Emmanuel II monument in Rome, containing the Altare della Patria, was built in his honor.
Biography
Victor Emmanuel was born as the eldest son of Charles Albert, Prince of Carignano, and Maria Theresa of Austria. His father succeeded a distant cousin as King of Sardinia-Piedmont in 1831. He lived for some years of his youth in Florence and showed an early interest in politics, the military, and sports. In 1842, he married his cousin, Adelaide of Austria. He was styled as the Duke of Savoy prior to becoming King of Sardinia-Piedmont.
He took part in the First Italian War of Independence (1848–1849) under his father, King Charles Albert, fighting in the front line at the battles of Pastrengo, Santa Lucia, Goito and Custoza.
He became King of Sardinia-Piedmont in 1849 when his father abdicated the throne, after being defeated by the Austrians at the Battle of Novara. Victor Emmanuel was immediately able to obtain a rather favorable armistice at Vignale by the Austrian imperial army commander, Radetzky. The treaty, however, was not ratified by the Piedmontese lower parliamentary house, the Chamber of Deputies, and Victor Emmanuel retaliated by firing his Prime Minister, Claudio Gabriele de Launay, replacing him with Massimo D'Azeglio. After new elections, the peace with Austria was accepted by the new Chamber of Deputies. In 1849, Victor Emmanuel also fiercely suppressed a revolt in Genoa, defining the rebels as a "vile and infected race of canailles."
In 1852, he appointed Count Camillo Benso of Cavour ("Count Cavour") as Prime Minister of Piedmont-Sardinia. This turned out to be a wise choice, since Cavour was a political mastermind and a major player in the Italian unification in his own right. Victor Emmanuel II soon became the symbol of the "Risorgimento", the Italian unification movement of the 1850s and early 60s. He was especially popular in the Kingdom of Sardinia-Piedmont because of his respect for the new constitution and his liberal reforms.
Crimean War
Following Victor Emmanuel's advice, Cavour joined Britain and France in the Crimean War against Russia. Cavour was reluctant to go to war due to the power of Russia at the time and the expense of doing so. Victor Emmanuel, however, was convinced of the rewards to be gained from the alliance created with Britain and, more importantly, France.
After successfully seeking British support and ingratiating himself with France and Napoleon III at the Congress of Paris in 1856 at the end of the war, Count Cavour arranged a secret meeting with the French emperor. In 1858, they met at Plombières-les-Bains (in Lorraine), where they agreed that if the French were to help Piedmont combat Austria, which still reigned over the Kingdom of Lombardy-Venetia in northern Italy, France would be awarded Nice and Savoy.
Wars of Italian Unification
The Italo-French campaign against Austria in 1859 started successfully. However, sickened by the casualties of the war and worried about the mobilisation of Prussian troops, Napoleon III secretly made a treaty with Franz Joseph of Austria at Villafranca whereby Piedmont would only gain Lombardy. France did not as a result receive the promised Nice and Savoy, but Austria did keep Venetia, a major setback for the Piedmontese, in no small part because the treaty had been prepared without their knowledge. After several quarrels about the outcome of the war, Cavour resigned, and the king had to find other advisors. France indeed only gained Nice and Savoy after the Treaty of Turin was signed in March 1860, after Cavour had been reinstalled as Prime Minister, and a deal with the French was struck for plebiscites to take place in the Central Italian Duchies.
Later that same year, Victor Emmanuel II sent his forces to fight the papal army at Castelfidardo and drove the Pope into Vatican City. His success at these goals led him to be excommunicated from the Catholic Church until 1878, when it was lifted just before his death. Then, Giuseppe Garibaldi conquered Sicily and Naples, and Sardinia-Piedmont grew even larger. On 17 March 1861 the Kingdom of Italy was officially established and Victor Emmanuel II became its king.
Victor Emmanuel supported Giuseppe Garibaldi's Expedition of the Thousand (1860–1861), which resulted in the rapid fall of the Kingdom of the Two Sicilies in southern Italy. However, the king halted Garibaldi when he appeared ready to attack Rome, still under the Papal States, as it was under French protection. In 1860, through local plebiscites, Tuscany, Modena, Parma and Romagna decided to side with Sardinia-Piedmont. Victor Emmanuel then marched victoriously in the Marche and Umbria after the victorious battle of Castelfidardo (1860) over the Papal forces.
The king subsequently met with Garibaldi at Teano, receiving from him the control of southern Italy. Another series of plebiscites in the occupied lands resulted in the proclamation of Victor Emmanuel as the first King of Italy by the new Parliament of unified Italy, on 17 March 1861. He did not renumber himself after assuming the new royal title, however. Turin became the capital of the new state. Only Rome, Veneto, and Trentino remained to be conquered.
Completion of the unification
In 1866 Victor Emmanuel allied himself with Prussia in the Third Italian War of Independence. Although not victorious in the Italian theater, he managed anyway to receive Veneto after the Austrian defeat in Germany. The British Foreign Secretary, Lord Clarendon, visited Florence in December 1867 and reported to London after talking to various Italian politicians: "There is universal agreement that Victor Emmanuel is an imbecile; he is a dishonest man who tells lies to everyone; at this rate he will end up losing his crown and ruining both Italy and his dynasty." In 1870, after two failed attempts by Garibaldi, he also took advantage of the Prussian victory over France in the Franco-Prussian War to capture Rome after the French withdrew. He entered Rome on 20 September 1870 and set up the new capital there on 2 July 1871, after a temporary move to Florence in 1864. The new Royal residence was the Quirinal Palace.
The rest of Victor Emmanuel II's reign was much quieter. After the Kingdom of Italy was established he decided to continue on as King Victor Emmanuel II instead of Victor Emmanuel I of Italy. This was a terrible move as far as public relations went as it was not indicative of the fresh start that the Italian people wanted and suggested that Sardinia-Piedmont had taken over the Italian Peninsula, rather than unifying it. Despite this mishap, the remainder of Victor Emmanuel II's reign was consumed by wrapping up loose ends and dealing with economic and cultural issues. His role in day-to-day governing gradually dwindled, as it became increasingly apparent that a king could no longer keep a government in office against the will of Parliament. As a result, while the wording of the Statuto Albertino stipulating that ministers were solely responsible to the crown remained unchanged, in practice they were now responsible to Parliament.
Victor Emmanuel died in Rome in 1878, after meeting with the envoys of Pope Pius IX, who had reversed the excommunication, and received last rites. He was buried in the Pantheon. His successor was his son Umberto I.
Family and children
In 1842 he married his first cousin once removed Adelaide of Austria (1822–1855). With her, he had eight children:
Maria Clotilde (1843–1911), who married Napoléon Joseph (the Prince Napoléon). Their grandson Prince Louis Napoléon was the Bonapartist pretender to the French imperial throne.
Umberto (1844–1900), later King of Italy. He married his first cousin Margherita of Savoy.
Amadeo (1845–1890), later King of Spain. He married Maria Vittoria dal Pozzo and later Maria Letizia Bonaparte.
Oddone Eugenio Maria (1846–1866), Duke of Montferrat.
Maria Pia (1847–1911), who married King Louis of Portugal.
Carlo Alberto (2 June 1851 – 28 June 1854), Duke of Chablais.
Vittorio Emanuele (6 July 1852 – 6 July 1852).
Vittorio Emanuele (18 January 1855 – 17 May 1855), Count of Geneva.
In 1869 he married morganatically his principal mistress Rosa Vercellana (3 June 1833 – 26 December 1885). Popularly known in Piedmontese as "Bela Rosin", she was born a commoner but made Countess of Mirafiori and Fontanafredda in 1858. Their offspring were:
Vittoria Guerrieri (2 December 1848 – 29 December 1905), married three times: to Giacomo Spinola, Luigi Spinola and Paolo DeSimone. She had issue.
Emanuele Alberto Guerrieri (16 March 1851 – 24 December 1894), Count of Mirafiori and Fontanafredda, married and had issue.
In addition to his morganatic second wife, Victor Emmanuel II had several other mistresses:
1) Laura Bon at Stupinigi, who bore him one daughter:
Emanuela Alberta Maria Vittoria Guerriero di Roverbella (6 September 1852 – 1890).
2) Unknown mistress at Mondovì,
mother of:
Donato Etna (1858-1938) who became a soldier during the First World War.
3) Virginia Rho at Turin, mother of two children:
Vittorio di Rho (1861 – Turin, 10 October 1913). He became a notable photographer.
Maria Pia di Rho (25 February 1866 – Vienna, 19 April 1947). Married to count Alessandro Montecuccoli.
4) Rosalinda Incoronata De Domenicis (1846-1916), mother of one daughter:
Vittoria De Domenicis (1869-1935) who married doctor Alberto Benedetti (1870-1920), with issue.
5) Baroness Vittoria Duplessis, who bore him:
Maria Savoiarda Guerrieri (1853-1885) that married Vincenzo Pietraforte (1853-1902).
6) Angela Rosa De Filippo, mother of:
Actor Domenico Scarpetta (1876-1952)
Honours
Italian
Knight of the Order of the Annunciation, 23 December 1836; Grand Master, 23 March 1849
Grand Cross of the Order of Saints Maurice and Lazarus, 1836; Grand Master, 23 March 1849
Grand Master of the Military Order of Savoy
Grand Master of the Order of the Crown of Italy
Grand Master of the Civil Order of Savoy
Gold Medal of Military Valour
Silver Medal of Military Valour
Medal of the Liberation of Rome (1849–1870)
Commemorative Medal of Campaigns of Independence Wars
Commemorative Medal of the Unity of Italy
Tuscan Grand Ducal family: Grand Cross of the Order of St. Joseph
Foreign
:
Knight of the Order of the Golden Fleece, 1841
Grand Cross of the Order of St. Stephen, 1869
:
Knight of the House Order of Fidelity, 1864
Grand Cross of the Order of the Zähringer Lion, 1864
: Knight of the Order of St. Hubert, 1869
: Grand Cordon of the Order of Leopold, 25 July 1855
: Knight of the Order of the Elephant, 2 September 1861
French Empire:
Médaille militaire
Commemorative medal of the 1859 Italian Campaign
: Grand Cross of the Order of Kamehameha I, 1865
Mexican Empire: Grand Cross of the Order of the Mexican Eagle, with Collar, 1865
Kingdom of Prussia:
Knight of the Order of the Black Eagle, 12 January 1866; with Collar, 1875
Pour le Mérite (military), 29 May 1872
: Knight of the Order of the Rue Crown, 1850
Sweden-Norway: Knight of the Order of the Seraphim, 30 August 1861
Beylik of Tunis: Husainid Family Order
: Stranger Knight of the Order of the Garter, 5 December 1855
Ancestry
See also
Unification of Italy
Giuseppe Garibaldi
Giuseppe Mazzini
Count Cavour
September Convention
Monument to Vittorio Emanuele II
List of famous big game hunters
Notes
References
Sources
old interpretations but useful on details; vol 1 goes to 1859]; volume 2 online covers 1859–62
In Italian
External links
External link: Genealogy of recent members of the House of Savoy
View of Venezia Square Victor Emmanuel II monument
1820 births
1878 deaths
19th-century kings of Italy
19th-century kings of Sardinia
Nobility from Turin
Grand Masters of the Gold Medal of Military Valor
Italian monarchs
Kings of Italy (1861–1946)
Princes of Savoy
Italian people of the Italian unification
Italian people of Polish descent
Italian generals
Claimant Kings of Jerusalem
Princes of Savoy-Carignan
Modern history of Italy
People of the Revolutions of 1848
Roman Catholic monarchs
People temporarily excommunicated by the Catholic Church
Burials at the Pantheon, Rome
Knights Grand Cross of the Order of Saints Maurice and Lazarus
Extra Knights Companion of the Garter
Knights of the Golden Fleece of Austria
Grand Crosses of the Order of Saint Stephen of Hungary
Recipients of the Pour le Mérite (military class)
People from Turin |
null | null | Debbie Reynolds | eng_Latn | Mary Frances "Debbie" Reynolds (April 1, 1932 – December 28, 2016) was an American actress, singer, and businesswoman. Her career spanned almost 70 years. She was nominated for the Golden Globe Award for Most Promising Newcomer for her portrayal of Helen Kane in the 1950 film Three Little Words. Her breakout role was her first leading role, as Kathy Selden in Singin' in the Rain (1952). Her other successes include The Affairs of Dobie Gillis (1953), Susan Slept Here (1954), Bundle of Joy (1956 Golden Globe nomination), The Catered Affair (1956 National Board of Review Best Supporting Actress Winner), and Tammy and the Bachelor (1957), in which her performance of the song "Tammy" reached number one on the Billboard music charts. In 1959, she released her first pop music album, titled Debbie.
She starred in Singin' in the Rain (1952), How the West Was Won (1962), and The Unsinkable Molly Brown (1964), a biographical film about the famously boisterous Molly Brown. Her performance as Brown earned her a nomination for the Academy Award for Best Actress. Her other films include The Singing Nun (1966), Divorce American Style (1967), What's the Matter with Helen? (1971), Charlotte's Web (1973), Mother (1996) (Golden Globe nomination), and In & Out (1997). Reynolds was also a cabaret performer. In 1979, she founded the Debbie Reynolds Dance Studio in North Hollywood, which was demolished in 2019 after being sold at auction, despite efforts to turn it into a museum.
In 1969, she starred on television in The Debbie Reynolds Show, for which she received a Golden Globe nomination. In 1973, Reynolds starred in a Broadway revival of the musical Irene and was nominated for the Tony Award for Best Lead Actress in a Musical. She was also nominated for a Daytime Emmy Award for her performance in A Gift of Love (1999) and an Emmy Award for playing Grace's mother Bobbi on Will & Grace. At the turn of the millennium, Reynolds reached a new, younger generation with her role as Aggie Cromwell in Disney's Halloweentown series. In 1988, she released her autobiography, titled Debbie: My Life. In 2013, she released a second autobiography, Unsinkable: A Memoir.
Reynolds also had several business ventures, including ownership of a dance studio and a Las Vegas hotel and casino, and she was an avid collector of film memorabilia, beginning with items purchased at the landmark 1970 MGM auction. She served as president of The Thalians, an organization dedicated to mental-health causes. Reynolds continued to perform successfully on stage, television, and film into her 80s. In January 2015, Reynolds received the Screen Actors Guild Life Achievement Award. In 2016, she received the Jean Hersholt Humanitarian Award. In the same year, a documentary about her life was released titled Bright Lights: Starring Carrie Fisher and Debbie Reynolds, which turned out to be her final film appearance; the film premiered on HBO on January 7, 2017.
Reynolds died following a hemorrhagic stroke on December 28, 2016, one day after the death of her daughter, fellow actress Carrie Fisher.
Early life
Reynolds was born on April 1, 1932, in El Paso, Texas, to Maxene N. "Minnie" Harman and Raymond Francis "Ray" Reynolds, a carpenter who worked for the Southern Pacific Railroad. She was of Scottish-Irish and English ancestry and was raised in a strict Nazarene church of her domineering mother. She had an older brother, William, who was two years her senior. Reynolds was a Girl Scout, once saying that she wanted to die as the world's oldest living Girl Scout. Reynolds was also a member of The International Order of Job's Daughters.
Her mother took in laundry for income, while they lived in a shack on Magnolia Street in El Paso. "We may have been poor," she said in a 1963 interview, "but we always had something to eat, even if Dad had to go out in the desert and shoot jackrabbits."
Her family moved to Burbank, California, in 1939. When Reynolds was a 16-year-old student at Burbank High School in 1948, she won the Miss Burbank beauty contest. Soon after, she was offered a contract with Warner Brothers and was given the stage name "Debbie" by studio head Jack L. Warner.
One of her closest high school friends said that she rarely dated during her teenaged years in Burbank.
Reynolds agreed, saying, "when I started, I didn't even know how to dress. I wore dungarees and a shirt. I had no money, no taste, and no training." Her friend adds:
Career
Film and television
Reynolds was discovered by talent scouts from Warner Bros. and MGM, who were at the 1948 Miss Burbank contest. Both companies wanted her to sign up with their studio, and had to flip a coin to see which one got her. Warner Bros. won the coin toss, and she was with the studio for two years. When Warner Bros. stopped producing musicals, she moved to MGM.
With MGM, Reynolds regularly appeared in movie musicals during the 1950s, and had several hit records during the period. Her song "Aba Daba Honeymoon" (featured in the film Two Weeks with Love (1950) and sung as a duet with co-star Carleton Carpenter) was the first soundtrack recording to become a top-of-the-chart gold record, reaching number three on the Billboard charts.
Her performance in the film greatly impressed the studio, which then gave her a co-starring role in what became her highest-profile film, Singin' in the Rain (1952), a satire on movie-making in Hollywood during the transition from silent to sound pictures. It co-starred Gene Kelly, whom she called a "great dancer and cinematic genius," adding, "He made me a star. I was 18 and he taught me how to dance and how to work hard and be dedicated." In 1956, she appeared in the musical Bundle of Joy with her then-husband, Eddie Fisher.
Reynolds was one of 14 top-billed names in How the West Was Won (1962) but she was the only one who appeared throughout, the story largely following the life and times of her character Lilith Prescott. In the film, she sang three songs: What Was Your Name in the States?, as her pioneering family begin their westward journey; Raise a Ruckus Tonight, starting a party around a wagon train camp fire; and, three times, Home in the Meadow – to the tune of Greensleeves with lyrics by Sammy Cahn.
Her starring role in The Unsinkable Molly Brown (1964) led to a nomination for the Academy Award for Best Actress. Reynolds noted that she initially had issues with its director, Charles Walters. "He didn't want me," she said. "He wanted Shirley MacLaine," who at the time was unable to take the role. "He said, 'You are totally wrong for the part." But six weeks into production, he reversed his opinion. "He came to me and said, "I have to admit that I was wrong. You are playing the role really well. I'm pleased." Reynolds also played in Goodbye Charlie, a 1964 comedy film about a callous womanizer who gets his just reward. It was adapted from George Axelrod's play Goodbye, Charlie and also starred Tony Curtis and Pat Boone.
She next portrayed Jeanine Deckers in The Singing Nun (1966). In what Reynolds once called the "stupidest mistake of my entire career", she made headlines in 1970 after instigating a fight with the NBC television network over cigarette advertising on her weekly television show. Although she was television's highest-paid female performer at the time, she quit the show for breaking its contract:
When NBC explained to Reynolds that banning cigarette commercials from her show would be impossible, she kept her resolve. The show drew mixed reviews, but according to NBC, it captured about 42% of the nation's viewing audience. She said later she was especially concerned about the commercials because of the number of children watching the show. She did quit doing the show after about a year, which she said had cost her about $2 million of lost income: "Maybe I was a fool to quit the show, but at least I was an honest fool. I'm not a phony or pretender. With me, it wasn't a question of money, but integrity. I'm the one who has to live with myself." The dispute would have been rendered moot and in Reynolds' favor anyway had she not resigned; by 1971, the Public Health Cigarette Smoking Act (which had been passed into law before she left the show) would ban all radio and television advertising for tobacco products.
Reynolds played the title role in the Hanna-Barbera animated musical Charlotte's Web (1973), in which she originated the song "Mother Earth and Father Time". Reynolds continued to make other appearances in film and television. She played Helen Chappel Hackett's mother, Deedee Chappel, on an episode of Wings titled, "If It's Not One Thing, It's Your Mother", which originally aired on November 22, 1994.
From 1999 to 2006, she played Grace Adler's theatrical mother, Bobbi Adler, on the NBC sitcom Will & Grace, which earned Reynolds her only Emmy Award nomination for Outstanding Guest Actress in a Comedy Series in 2000. She played a recurring role in the Disney Channel Original Movie Halloweentown film series as Aggie Cromwell. Reynolds made a guest appearance as a presenter at the 69th Academy Awards in 1997.
In 2000, Reynolds took up a recurring voice role on the children's television program Rugrats, playing the grandmother of two of the characters. In 2001, she co-starred with Elizabeth Taylor, Shirley MacLaine, and Joan Collins in the comedy These Old Broads, a television movie written for her by her daughter, Carrie Fisher. She had a cameo role as herself in the 2004 film Connie and Carla. In 2013, she appeared in Behind the Candelabra, as the mother of Liberace.
Reynolds appears with her daughter in Bright Lights: Starring Carrie Fisher and Debbie Reynolds, a 2016 documentary about the very close relationship between the two. It premiered at the 2016 Cannes Film Festival. The television premiere was January 7, 2017, on HBO. According to USA Today, the film is "an intimate portrait of Hollywood royalty ... [it] loosely chronicles their lives through interviews, photos, footage, and vintage home movies... It culminates in a moving scene, just as Reynolds is preparing to receive the 2015 Screen Actors Guild Life Achievement Award, which Fisher presented to her mom."
Music career and cabaret
Her recording of the song "Tammy" (1957; from Tammy and the Bachelor), earned her a gold record. It was a number one single on the Billboard pop charts in 1957. In the movie (the first of the Tammy film series), she co-starred with Leslie Nielsen.
Reynolds also scored two other top-25 Billboard hits with "A Very Special Love" (number 20 in January 1958) and "Am I That Easy to Forget" (number 25 in March 1960)—a pop-music version of a country-music hit made famous by Carl Belew (in 1959), Skeeter Davis (in 1960), and several years later by singer Engelbert Humperdinck.
In 1991, she released an album titled The Best of Debbie Reynolds.
For 10 years, she headlined for about three months a year in Las Vegas's Riviera Hotel. She enjoyed live shows, though that type of performing "was extremely strenuous," she said.
With a performing schedule of two shows a night, seven nights a week, it's probably the toughest kind of show business, but in my opinion, the most rewarding. I like the feeling of being able to change stage bits and business when I want. You can't do that in motion pictures or TV.
As part of her nightclub act, Reynolds was noted for doing impressions of celebrities such as Eva and Zsa Zsa Gabor, Mae West, Barbra Streisand, Phyllis Diller, and Bette Davis. Her impersonation of Davis was inspired following their co-starring roles in the 1956 film, The Catered Affair. Reynolds had started doing stage impersonations as a teenager; her impersonation of Betty Hutton was performed as a singing number during the Miss Burbank contest in 1948.
Reynolds' last album was a Christmas record with Donald O'Connor entitled Chrissy the Christmas Mouse arranged and conducted by Angelo DiPippo.
Reynolds was also a French horn player. Gene Kelly, reflecting on Reynolds's sudden fame, recalled, "There were times when Debbie was more interested in playing the French horn somewhere in the San Fernando Valley or attending a Girl Scout meeting....She didn’t realize she was a movie star all of a sudden."
Stage work
With limited film and television opportunities coming her way, Reynolds accepted an opportunity to make her Broadway debut. She starred in the 1973 revival of Irene, a musical first produced 60 years before. When asked why she waited so long to appear in a Broadway play, she explained:
Reynolds and her daughter Carrie both made their Broadway debuts in the play. Per reports, the production broke records for the highest weekly gross of any musical. For that production, she received a Tony nomination. Reynolds also starred in a self-titled Broadway revue, Debbie, in 1976. She toured with Harve Presnell in Annie Get Your Gun, then wrapped up the Broadway run of Woman of the Year in 1983. In the late 1980s, Reynolds repeated her role as Molly Brown in the stage version of The Unsinkable Molly Brown, first opposite Presnell (repeating his original Broadway and movie role) and later with Ron Raines.
Best Foot Forward (1953) (Dallas State Fair)
Irene (1973) (Broadway and US national tour)
Debbie (1976) (Broadway)
Annie Get Your Gun (1977) (San Francisco and Los Angeles)
Woman of the Year (1982) (Broadway) (replacement for Lauren Bacall)
The Unsinkable Molly Brown (1989) (US national tour)
Irene (2008) Perth Western Australia
In 2010, she appeared in her own West End show Debbie Reynolds: Alive and Fabulous.
Film history preservation
Reynolds amassed a large collection of movie memorabilia, beginning with items from the landmark 1970 Metro-Goldwyn-Mayer auction, and she displayed them, first in a museum at her Las Vegas hotel and casino during the 1990s and later in a museum close to the Kodak Theater in Los Angeles.
The museum was to relocate to be the centerpiece of the Belle Island Village tourist attraction in the resort city of Pigeon Forge, Tennessee, but the developer went bankrupt. The museum filed for Chapter 11 bankruptcy in June 2009. The most valuable asset of the museum was Reynolds' collection. Todd Fisher, Reynolds' son, announced that his mother was "heartbroken" to have to auction off the collection. It was valued at $10.79 million in the bankruptcy filing. Los Angeles auction firm Profiles in History was given the responsibility of conducting a series of auctions. Among the "more than 3500 costumes, 20,000 photographs, and thousands of movie posters, costume sketches, and props" included in the sales were Charlie Chaplin's bowler hat and Marilyn Monroe's white "subway dress", whose skirt is lifted up by the breeze from a passing subway train in the film The Seven Year Itch (1955). The dress sold for $4.6 million in 2011; the final auction was held in May 2014.
Business ventures
In 1979, Reynolds opened her own dance studio in North Hollywood. In 1983, she released an exercise video, Do It Debbie's Way!. She purchased the Clarion Hotel and Casino, a hotel and casino in Las Vegas, in 1992. She renamed it the Debbie Reynolds Hollywood Hotel. It was not a success. In 1997, Reynolds was forced to declare bankruptcy. In June 2010, she replaced Ivana Trump answering reader queries for the weekly paper Globe.
Marriages and later life
Reynolds was married three times. Her first marriage was to singer Eddie Fisher in 1955. They became the parents of Carrie Fisher and Todd Fisher. The couple divorced in 1959 when it was revealed shortly after the death of Elizabeth Taylor's husband Mike Todd that Fisher had been having an affair with her; Taylor and Reynolds were good friends at the time. The Eddie FisherElizabeth Taylor affair was a great public scandal, which led to the cancellation of Eddie Fisher's television show.
In 2011, Reynolds was on The Oprah Winfrey Show just weeks before Elizabeth Taylor's death. She explained that Taylor and she happened to be traveling at the same time on the ocean liner Queen Elizabeth some time in the late 1960s or early 1970s, when they reconciled. Reynolds sent a note to Taylor's room, and Taylor sent a note in reply asking to have dinner with Reynolds and end their feud. As Reynolds described it, "we had a wonderful evening with a lot of laughs." In 1972, she noted the bright side of the divorce and her remarriage:
Reynolds' second marriage, to millionaire businessman Harry Karl, lasted from 1960 to 1973. For a period during the 1960s, she stopped working at the studio on Friday afternoons to attend Girl Scout meetings, since she was the leader of the Girl Scout Troop of which her 13-year-old daughter Carrie and her stepdaughter Tina Karl, also 13, were members. Reynolds later found herself in financial difficulty because of Karl's gambling and bad investments. Reynolds' third marriage was to real estate developer and con man Richard Hamlett from 1984 to 1996.
In 2011, Reynolds stepped down after 56 years of involvement in The Thalians, a charitable organization devoted to children and adults with mental-health issues.
Reynolds was hospitalized in October 2012 at Cedars-Sinai Medical Center in Los Angeles due to an adverse reaction to medication. She cancelled appearances and concert engagements for the next three months.
Death and legacy
On December 23, 2016, Reynolds's daughter, actress and writer Carrie Fisher, suffered a medical emergency on a transatlantic flight from London to Los Angeles, and died on December 27, 2016, at the age of 60 at Ronald Reagan UCLA Medical Center. The following day, December 28, Reynolds was taken by ambulance to Cedars-Sinai Medical Center in Los Angeles, after suffering a "severe stroke," according to her son. Later that afternoon, Reynolds was pronounced dead in the hospital; she was 84 years old. On January 9, 2017, her cause of death was determined to be an intracerebral hemorrhage, with hypertension a contributing factor.
Todd Fisher later said that Reynolds had been seriously affected by her daughter's death, and that her grief was partially responsible for her stroke, noting that his mother had stated, "I want to be with Carrie", shortly before she died. During an interview for the December 30, 2016 airing of the ABC-TV program 20/20, Fisher elaborated on this, saying that his mother had joined his sister in death because Reynolds "didn't want to leave Carrie and did not want her to be alone". He added, "she didn't die of a broken heart" as some news reports had implied, but rather "just left to be with Carrie".
Reynolds was entombed, while her daughter was cremated. A portion of Carrie Fisher's ashes was laid to rest beside Reynolds' crypt at Forest Lawn Memorial Park in Hollywood Hills during a larger joint memorial service held on March 25, while the remainder of Fisher's ashes are held in a giant, novelty Prozac pill.
Awards and honors
Reynolds was the 1955 Hasty Pudding Woman of the Year. Her footprints and handprints are preserved at Grauman's Chinese Theatre in Hollywood, California. She also has a star on the Hollywood Walk of Fame, at 6654 Hollywood Boulevard, for live performance and a Golden Palm Star on the Palm Springs, California, Walk of Stars dedicated to her. In keeping with the celebrity tradition of the Shenandoah Apple Blossom Festival of Winchester, Virginia, Reynolds was honored as the Grand Marshal of the 2011 ABF that took place from April 26 to May 1, 2011.
On November 4, 2006, Reynolds received the Lifetime Achievement in the Arts Award from Chapman University (Orange, California). On May 17, 2007, she was awarded an honorary degree of Doctor of Humane Letters from the University of Nevada, Reno, where she had contributed for many years to the film studies program.
Filmography
Short subjects
A Visit with Debbie Reynolds (1959)
The Story of a Dress (1964)
In the Picture (2012)
Partial television credits
Radio broadcasts
See also
List of American film actresses
List of people from California
List of people from Texas
References
Further reading
External links
Debbie Reynolds at TVGuide.com
Photographs and literature
Debbie Reynolds at Emmys.com
The Official Academy Awards Database: Type "Debbie Reynolds" at the Nominee box
Debbie Reynolds profile at Aveleyman.com
1932 births
2016 deaths
20th-century American actresses
20th-century American singers
20th-century American women singers
20th-century American women writers
21st-century American actresses
21st-century American non-fiction writers
21st-century American singers
21st-century American women singers
21st-century American women writers
Actresses from Burbank, California
Actresses from El Paso, Texas
American autobiographers
American beauty pageant winners
American collectors
American female dancers
American women pop singers
American film actresses
American impressionists (entertainers)
American members of the Church of the Nazarene
American musical theatre actresses
American people of English descent
American people of Scotch-Irish descent
American stage actresses
American tap dancers
American television actresses
American voice actresses
American women non-fiction writers
Burials at Forest Lawn Memorial Park (Hollywood Hills)
Carrie Fisher
Dancers from California
Dancers from Texas
Deaths by intracerebral hemorrhage
Deaths from hypertension
Film memorabilia
Jean Hersholt Humanitarian Award winners
Metro-Goldwyn-Mayer contract players
MGM Records artists
Musicians from Burbank, California
Musicians from El Paso, Texas
Neurological disease deaths in California
Traditional pop music singers
Warner Bros. contract players
Western (genre) film actresses
Women autobiographers
Women collectors
Writers from California
Writers from Texas |
null | null | Kiss Me, Kate | eng_Latn | Kiss Me, Kate is a musical written by Bella and Samuel Spewack with music and lyrics by Cole Porter. The story involves the production of a musical version of William Shakespeare's The Taming of the Shrew and the conflict on and off-stage between Fred Graham, the show's director, producer, and star, and his leading lady, his ex-wife Lilli Vanessi. A secondary romance concerns Lois Lane, the actress playing Bianca, and her gambler boyfriend, Bill, who runs afoul of some gangsters. The original production starred Alfred Drake, Patricia Morison, Lisa Kirk and Harold Lang.
Kiss Me, Kate was Porter's response to Rodgers and Hammerstein's Oklahoma! and other integrated musicals; it was the first show he wrote in which the music and lyrics were firmly connected to the script. The musical premiered in 1948 and proved to be Porter's only show to run for more than 1,000 performances on Broadway. In 1949, it won the first Tony Award for Best Musical.
Inspiration
The musical was inspired by the on-stage/off-stage battling of husband-and-wife actors Alfred Lunt and Lynn Fontanne during their 1935 production of The Taming of the Shrew, witnessed by future Broadway producer Arnold Saint-Subber. In 1947, he asked the Spewacks (undergoing their own marital woes at the time) to write the script; Bella Spewack in turn enlisted Cole Porter to write the music and lyrics.
Productions
Original Broadway production
After a 3½-week pre-Broadway tryout at the Shubert Theatre in Philadelphia starting December 2, 1948, the original Broadway production opened on December 30, 1948, at the New Century Theatre, where it ran for nineteen months before transferring to the Shubert, for a total run of 1,077 performances. Directed by John C. Wilson with choreography by Hanya Holm, the original cast included Alfred Drake, Patricia Morison, Lisa Kirk, Harold Lang, Charles Wood and Harry Clark. The 1949 original cast recording has been inducted into the Library of Congress's National Recording Registry for the album's "cultural, artistic and/or historical significance to American society and the nation's audio legacy".
Original London and Australian productions
The original West End production opened on March 8, 1951, at the Coliseum Theatre, and ran for 400 performances. Directed by Sam Spewack with choreography again by Holm, this production starred Patricia Morison, Bill Johnson, Adelaide Hall and Julie Wilson.
The original Australian production played from February 1952 at His Majesty's Theatre, Melbourne, before seasons in Sydney, Brisbane, Perth and Adelaide until 1954. The production featured Hayes Gordon and Joy Turpin (later replaced by Evie Hayes).
1970 London and 1987 UK revivals
A London revival opened in December 1970 at the London Coliseum, in a production by the Sadler's Wells Opera. The cast featured Emile Belcourt (Petruchio), Judith Bruce, Eric Shilling, Ann Howard (Kate), Francis Egerton and Robert Lloyd, with direction by Peter Coe and choreography by Sheila O'Neill. Coe did a translation for British audiences, including having "a tea wagon", and included "traditional English music hall jokes". This revival had a "brief run", according to the Encyclopedia of the Musical Theatre.
The Royal Shakespeare Company staged a production which opened at the Royal Shakespeare Theatre, Stratford-upon-Avon, on February 10, 1987, toured the UK from March to May, and then played at London's Old Vic Theatre from May 19, 1987. Directed by Adrian Noble and staged by Ron Field, the production starred Nichola McAuliffe and Paul Jones as Lilli/Kate and Fred/Petruchio, with Tim Flavin and Fiona Hendley as Bill/Lucentio and Lois/Bianca. The gangsters were played by Emil Wolk and John Bardon, who shared the 1987 Olivier Award for Outstanding Performance of the Year by an Actor in a Musical, while McAuliffe won the Olivier for Outstanding Performance of the Year by an Actress in a Musical. The production moved to the Savoy Theatre on January 15, 1988, with a new cast.
1952 and 1999 Broadway revivals
A short-lived Broadway revival ran at the Broadway Theatre in January 1952. It was directed by John C. Wilson and choreographed by Hanya Holm. Holly Harris and Robert Wright starred as Lilli and Fred.
A Broadway revival opened at the Martin Beck Theatre on November 18, 1999, and closed on December 30, 2001, after 881 performances and 28 previews. Directed by Michael Blakemore, Produced by Richard Godwin, and choreographed by Kathleen Marshall and Rob Ashford, the opening night cast included Marin Mazzie, Brian Stokes Mitchell, Amy Spanger, Michael Berresse, Ron Holgate, Lee Wilkof and Michael Mulheren. This production won the Tony Awards for Best Revival of a Musical and Best Actor in a Musical for Mitchell; Marin Mazzie received a Tony nomination for Best Actress in a Musical, and Michael Berresse, Lee Wilkof and Michael Mulheren received Tony nominations for Best Featured Actor in a Musical.
2001 and 2012 London revivals
A West End revival opened at the Victoria Palace Theatre on October 30, 2001, and closed on August 24, 2002. As with the 1999 Broadway revival, Michael Blakemore was the director with choreography by Kathleen Marshall. Brent Barrett and Marin Mazzie co-starred.
Chichester Festival Theatre's 2012 revival of the show transferred to the Old Vic Theatre on London's South Bank in November 2012, with an official opening in December. It starred Hannah Waddingham as Lili/Kate and Alex Bourne as Fred Graham. The production was directed by Trevor Nunn. The show received positive reviews from critics and audiences. Hannah Waddingham and Alex Bourne were both nominated for the 2013 Olivier Awards as Best Actress/Actor in a Musical for their performances.
2015 Opera North production
In September 2015 Opera North presented a revival directed by Jo Davies, choreographed by Will Tuckett. The production opened at the Leeds Grand Theatre before touring to Theatre Royal Newcastle, The Lowry Salford, and Theatre Royal Nottingham. The production was co-produced with Welsh National Opera who continued to tour it in 2016, first as part of the Shakespeare400 season at the Wales Millennium Centre, Cardiff, and then to Liverpool Empire Theatre, Bristol Hippodrome, New Theatre Oxford, Mayflower Theatre Southampton, Birmingham Hippodrome, and Venue Cymru Llandudno, before returning to Cardiff. The production was revived at London Coliseum briefly in June 2018, following a return visit to Leeds Grand Theatre in May 2018.
2019 Broadway revival
The Roundabout Theatre Company presented a revised third Broadway revival of Kiss Me, Kate with Kelli O'Hara as Lilli Vanessi/Kate, Will Chase as Fred Graham/Petruchio, Stephanie Styles as Lois Lane/Bianca, and Corbin Bleu as Bill Calhoun/Lucentio. The production is directed by Scott Ellis and choreographed by Warren Carlyle, with minor "feminist" updates by Amanda Green to make the musical "more accessible for today’s audiences". Roundabout produced a benefit concert of the show with O'Hara, Ellis, and Carlyle in 2016. The revival had a limited run at Studio 54. Previews began on February 14, 2019, with the opening on March 14. The production closed on June 30, 2019.
Casts
Original casts for long-running major productions
Plot
Act I
The cast of a musical version of William Shakespeare's The Taming of the Shrew is rehearsing for the opening of the show that evening ("Another Op'nin', Another Show"). Egotistical Fred Graham is the director and producer and is starring as Petruchio, and his movie-star ex-wife, Lilli Vanessi, is playing Katherine. The two seem to be constantly arguing, and Lilli is particularly angry that Fred is pursuing the sexy young actress Lois Lane, who is playing Bianca. After the rehearsal, Lois's boyfriend Bill appears; he is playing Lucentio, but he missed the rehearsal because he was gambling. He tells her that he signed a $10,000 IOU in Fred's name, and Lois reprimands him ("Why Can't You Behave?").
Before the opening, Fred and Lilli meet backstage, and Lilli shows off her engagement ring from Washington insider Harrison Howell, reminding Fred that it's the anniversary of their divorce. They recall the operetta in which they met, which included "Wunderbar", a Viennese waltz; they end up fondly reminiscing and singing and dancing. Two gangsters show up to collect the $10,000 IOU, and Fred replies that he never signed it. The gangsters obligingly say they will give him time to remember it and will return later. In her dressing room, Lilli receives flowers from Fred, and she declares that she is still "So In Love" with him. Fred tries to keep Lilli from reading the card that came with the flowers, which reveals that he really intended them for Lois. However, Lilli takes the card with her onstage, saying she will read it later.
The show begins ("We Open in Venice"). Baptista, Katherine and Bianca's father, will not allow his younger daughter Bianca to marry until his older daughter Katherine is married. However, she is shrewish and ill-tempered, and no man desires to marry her. Three suitors – Lucentio, Hortensio, and Gremio – try to woo Bianca, and she says that she would marry any of them ("Tom, Dick, or Harry"). Petruchio, a friend of Lucentio, expresses a desire to marry into wealth ("I've Come to Wive it Wealthily in Padua"). The suitors hatch a plan for him to marry Kate, as Baptista is rich. Kate, however, has no intentions of getting married ("I Hate Men"). Petruchio attempts to woo her ("Were Thine That Special Face"). Offstage, Lilli has an opportunity to read the card. She walks on stage off-cue and begins hitting Fred, who, along with the other actors, tries to remain in character as Baptista gives Petruchio permission to marry Kate. Lilli continues to strike Fred, and he ends up spanking her. Offstage, Lilli furiously declares she is leaving the show. However, the gangsters have reappeared, and Fred tells them that if Lilli quits, he'll have to close the show and won't be able to pay them the $10,000. The gangsters force her to stay at gunpoint. Back onstage, Bianca and Lucentio dance while the chorus performs "We Sing of Love", covering a scene change. The curtain opens, revealing the exterior of a church; Petruchio and Kate have just been married, and they exit the church; the gangsters, dressed in Shakespearean costume, are onstage to make sure that Lilli stays. Petruchio implores for Kate to kiss him, and she refuses. He lifts her over his shoulder and carries her offstage while she pummels his shoulder with her fists ("Kiss Me Kate").
Act II
During the show's intermission, the cast and crew relax in the alley behind the theater. Paul (Fred's assistant), along with a couple other crew members, lament that it's "Too Darn Hot" to meet their lovers that night. The play continues, and Petruchio tries to 'tame' Katherine and mourns for his now-lost bachelor life ("Where Is the Life That Late I Led?"). Off-stage, Lilli's fiancé Harrison Howell is looking for Lilli. He runs into Lois, and she recognizes him as a former lover but promises not to tell Lilli. Bill is shocked to overhear this, but Lois tells him that even if she is involved with other men, she is faithful to him in her own way ("Always True to You in My Fashion"). Lilli tries to explain to Howell that she is being forced to stay at the theatre by the gangsters, but Howell doesn't believe her and wants to discuss wedding plans. Fred insidiously points out how boring Lilli's life with Howell will be compared with the theatre. Bill sings a love song he has written for Lois ("Bianca").
The gangsters discover that their boss has been killed, so the IOU is no longer valid. Lilli leaves—without Howell—as Fred unsuccessfully tries to persuade her to stay ("So in Love" (Reprise)). The gangsters get caught on stage and improvise a tribute to Shakespeare in which they explain that knowing Shakespeare is the key to romance ("Brush Up Your Shakespeare"). The company prepares for the conclusion of the play, the wedding of Bianca and Lucentio, even though they are now missing one of the main characters. However, just in time for Katherine's final speech, Lilli arrives onstage ("I Am Ashamed That Women Are So Simple"). Fred and Lilli wordlessly reconcile on stage, and the play ends ("Kiss Me Kate" (Finale)) with them, as well as Bill and Lois, kissing passionately.
Song list
Act I
"Another Op'nin', Another Show" – Hattie and Company
"Why Can't You Behave?" – Lois Lane and Bill Calhoun
"Wunderbar" – Fred Graham and Lilli Vanessi
"So in Love" – Lilli Vanessi
"We Open in Venice" – Fred Graham, Lilli Vanessi, Lois Lane and Bill Calhoun
"Tom, Dick or Harry" – Bianca, Lucentio, Gremio and Hortensio
"I've Come to Wive it Wealthily in Padua" – Fred Graham and The Men
"I Hate Men" – Lilli Vanessi
"Were Thine That Special Face" – Fred Graham
"We Sing of Love (Cantiamo D'Amore)" – Lois Lane, Bill Calhoun and Ensemble
"Kiss Me, Kate" – Fred Graham, Lili Vanessi and Ensemble
Act II
"Too Darn Hot" – Paul and Ensemble
"Where is the Life That Late I Led?" – Fred Graham
"Always True to You in My Fashion" – Lois Lane
"From This Moment On" – Harrison Howell and Lilli Vanessi
"Bianca" – Bill Calhoun and Company
"So in Love" (Reprise) – Fred Graham
"Brush Up Your Shakespeare" – First Gangster and Second Gangster
"I Am Ashamed That Women Are So Simple" – Lilli Vanessi
"Kiss Me, Kate" (Reprise = Finale) – Company
Notes
In 1998, the original cast recording of the 1948 Broadway production was inducted into the Grammy Hall of Fame.
For the 1999 revival, "From This Moment On", originally from Porter's 1951 musical Out of This World, was added to Act II between "Always True to You in My Fashion" and "Bianca"; this song had been added to the 1953 film version of Kiss Me Kate.
For the 2019 revival, "I Am Ashamed That Women Are So Simple" was changed to "I Am Ashamed That People Are So Simple."
Film and television
A film version of the same name was released in 1953. There have been at least four television productions, the first on Hallmark Hall of Fame in 1958, with Drake and Morison reprising their Broadway roles, the second recorded for the launch of BBC Two in the UK in 1964, starring Howard Keel, Patricia Morison and Millicent Martin, the third in 1968 with then husband-and-wife team Robert Goulet and Carol Lawrence, and the fourth in 2003 on Great Performances, a high-definition shot performance of the London revival with Brent Barrett and Rachel York.
Awards and nominations
Original Broadway production
1987 Royal Shakespeare Company UK revival
1999 Broadway revival
2001 London revival
2012 London revival
2019 Broadway revival
Notes
References
McBrien, William (2000). Cole Porter. Reprint edition. New York: Knopf.
Stempel, Larry (2010). Showtime: A History of the Broadway Musical Theater. New York: W.W. Norton and Company.
Information from the TamsWitmark website
Information from the Imagination website
External links
Porter information
PBS article
Review of original Broadway production at Variety
2014 Critical Edition Orchestral score, Vocal score
Musicals by Cole Porter
1948 musicals
Broadway musicals
Drama Desk Award-winning musicals
Modern adaptations of works by William Shakespeare
Plays and musicals based on The Taming of the Shrew
West End musicals
Grammy Hall of Fame Award recipients
United States National Recording Registry recordings
American plays adapted into films
Tony Award for Best Musical
Tony Award-winning musicals
Backstage musicals |
null | null | List of winners and shortlisted authors of the Booker Prize | eng_Latn | The following is a list of winners and shortlisted authors of the Booker Prize for Fiction. Winning titles are listed in yellow, first in their year.
The prize has been awarded each year since 1969 to the best original full-length novel, written in the English language, by a citizen of the Commonwealth of Nations or the Republic of Ireland. In 2014, it was opened for the first time to any work published in the United Kingdom and written in (not translated into) the English language.
There have been three special awards celebrating the Booker's history. In 1993, the "Booker of Bookers" prize was awarded to Salman Rushdie for Midnight's Children (the 1981 winner) as the best novel to win the award in its first 25 years. Midnight's Children also won a public vote in 2008, on the prize's fortieth anniversary, "The Best of the Booker". In 2018 a special "Golden Booker" was awarded celebrating 50 years of the award - this was won by Michael Ondaatje for The English Patient.
Shortlists
Writers with multiple awards
2 Awards
J. M. Coetzee
Peter Carey
J. G. Farrell
Hilary Mantel
Margaret Atwood
Writers with multiple nominations
The following writers have received two or more nominations:
6 Nominations
Margaret Atwood
Iris Murdoch
5 Nominations
Beryl Bainbridge
Ian McEwan
Salman Rushdie
4 Nominations
Julian Barnes
Peter Carey
Penelope Fitzgerald
Kazuo Ishiguro
Thomas Keneally
Ali Smith
William Trevor
3 Nominations
Kingsley Amis
Sebastian Barry
J. M. Coetzee
Anita Desai
Damon Galgut
Doris Lessing
Penelope Lively
Hilary Mantel
Rohinton Mistry
Timothy Mo
Brian Moore
Muriel Spark
Colm Toibin
Barry Unsworth
Sarah Waters
2 Nominations
Paul Bailey
John Banville
Nina Bawden
Andre Brink
A. S. Byatt
J. L. Carr
Jim Crace
Roddy Doyle
Esi Edugyan
J. G. Farrell
Mohsin Hamid
Alan Hollinghurst
Howard Jacobson
James Kelman
Deborah Levy
David Lodge
Patrick McCabe
Tom McCarthy
David Mitchell
V. S. Naipaul
Chigozie Obioma
Richard Powers
Julian Rathbone
Mordecai Richler
Bernice Rubens
Carol Shields
David Storey
Graham Swift
Tim Winton
Notes
References
External links
Full details of the winners, judges and shortlisted books for all the Booker prizes (1969–2008), The Guardian, 10 October 2008.
Booker Prize for Fiction
+List
Booker Prize Winners |
null | null | Leading-tone | eng_Latn | In music theory, a leading-tone (also called a subsemitone, and a leading-note in the UK) is a note or pitch which resolves or "leads" to a note one semitone higher or lower, being a lower and upper leading-tone, respectively. Typically, the leading tone refers to the seventh scale degree of a major scale (), a major seventh above the tonic. In the movable do solfège system, the leading-tone is sung as ti.
A leading-tone triad is a triad built on the seventh scale degree in a major key (vii in Roman numeral analysis), while a leading-tone seventh chord is a seventh chord built on the seventh scale degree (vii7). Walter Piston considers and notates vii as V, an incomplete dominant seventh chord. (For the Roman numeral notation of these chords, see Roman numeral analysis.)
Note
Seventh scale degree (or lower leading tone)
Typically, when people speak of the leading tone, they mean the seventh scale degree () of the major scale, which has a strong affinity for and leads melodically to the tonic. It is sung as ti in movable-do solfège. For example, in the C major scale, the leading note is the note B.
As a diatonic function, the leading tone is the seventh scale degree of any diatonic scale when the distance between it and the tonic is a single semitone. In diatonic scales in which there is a whole tone between the seventh scale degree and the tonic, such as the Mixolydian mode, the seventh degree is called the subtonic. However, in modes without a leading tone, such as Dorian and Mixolydian, a raised seventh is often featured during cadences, such as in the harmonic minor scale.
A leading tone outside of the current scale is called a secondary leading-tone, leading to a secondary tonic. It functions to briefly tonicize a scale tone (usually the 5th degree) as part of a secondary dominant chord. In the second measure of Beethoven's Waldstein Sonata (shown below), the F's function as secondary leading-tones, which resolve to G in the next measure.
Descending, or upper, leading-tone
By contrast, a descending, or upper, leading-tone is a leading tone that resolves down, as opposed to the seventh scale degree (a lower leading-tone) which resolves up. The descending, or upper, leading-tone usually is a lowered second degree () resolving to the tonic, but the expression may at times refer to a resolving to the dominant. In German, the term Gegenleitton ("counter leading tone") is used by Hugo Riemann to denote the descending or upper leading-tone (), but Heinrich Schenker uses abwärtssteigenden Leitton or abwärts Leitton ("descending leading tone") to mean the descending diatonic supertonic ().)
The tritone substitution, chord progression ii-subV-I on C (Dm-Db7-C), results in an upper leading-note.
Analysis
According to Ernst Kurth, the major and minor thirds contain "latent" tendencies towards the perfect fourth and whole tone, respectively, and thus establish tonality. However, Carl contests Kurth's position, holding that this drive is in fact created through or with harmonic function, a root progression in another voice by a whole-tone or fifth, or melodically (monophonically) by the context of the scale. For example, the leading tone of alternating C chord and F minor chords is either the note E leading to F (if F is tonic), or A leading to G (if C is tonic).
In works from the 14th- and 15th-century Western tradition, the leading tone is created by the progression from imperfect to perfect consonances, such as a major third to a perfect fifth or minor third to a unison. The same pitch outside of the imperfect consonance is not a leading tone.
Forte claims that the leading tone is only one example of a more general tendency: the strongest progressions, melodic and harmonic, are by half step. He suggests that one play a G major scale and stop on the seventh note (F) to personally experience the feeling of lack caused by the "particularly strong attraction" of the seventh note to the eighth (F→G'), thus its name.
Leading-tone triad
A leading-tone chord is a triad built on the seventh scale degree in major and the raised seventh-scale-degree in minor. The quality of the leading-tone triad is diminished in both major and minor keys. For example, in both C major and C minor, it is a B diminished triad (though it is usually written in first inversion, as described below).
According to John Bunyan Herbert, (who uses the term "subtonic", which later came to usually refer to a seventh scale degree pitched a whole tone below the tonic note),
Function
The leading-tone triad is used in several functions. It is commonly used as a passing chord between a root position tonic triad and a first inversion tonic triad: that is, "In addition to its basic function of passing between I and I, VII has another important function: it can form a neighboring chord to I or I." In that instance, the leading-tone triad prolongs tonic through neighbor and passing motion. The example below shows two measures from the fourth movement of Beethoven's Piano Sonata No. 3 in C major, Op. 2 in which a leading-tone triad functions as a passing chord between I and I.
The leading-tone triad may also be regarded as an incomplete dominant seventh chord: "A chord is called 'Incomplete' when its root is omitted. This omission occurs, occasionally, in the chord of the dom.-seventh, and the result is a triad upon the leading-tone."
Some sources say the chord is not a chord; some argue it is an incomplete dominant seventh chord, especially when the diminished triad is written in its first inversion (resembling a second inversion dominant seventh without a root):
For example, vii often substitutes for V, which it closely resembles, and its use may be required in situations by voice leading: "In a strict four-voice texture, if the bass is doubled by the soprano, the VII [vii] is required as a substitute for the V".
Voice-leading
Since the leading-tone triad is a diminished triad, it is usually found in its first inversion: According to Carl Edward Gardner, "The first inversion of the triad is considered, by many, preferable to root position. The second inversion of the triad is unusual. Some theorists forbid its use."
In a four-part chorale texture, the third of the leading-tone triad is doubled in order to avoid adding emphasis on the tritone created by the root and the fifth. Unlike a dominant chord where the leading-tone can be frustrated and not resolve to the tonic if it is in an inner voice, the leading-tone in a leading-tone triad must resolve to the tonic. Commonly, the fifth of the triad resolves down since it is phenomenologically similar to the seventh in a dominant seventh chord. All in all, the tritone resolves inward if it is written as a diminished fifth (m. 1 below) and outward if it is written as an augmented fourth (m. 2).
Leading-tone seventh chord
The leading-tone seventh chords are vii7 and vii7, the half-diminished and diminished seventh chords on the seventh scale degree () of the major and harmonic minor. For example, in C major and C minor, the leading-tone seventh chords are B half-diminished (BDFA) and B diminished (B()DFA), respectively.
Leading-tone seventh chords were not characteristic of Renaissance music but are typical of the Baroque and Classical period. They are used more freely in Romantic music but began to be used less in classical music as conventions of tonality broke down. They are integral to ragtime and contemporary popular and jazz music genres.
Composers throughout the common practice period often employed modal mixture when using the leading-tone seventh chord in a major key, allowing for the substitution of the half-diminished seventh chord with the fully diminished seventh chord (by lowering its seventh). This mixture is commonly used when the leading-tone seventh chord is functioning as a secondary leading-tone chord.
The example below shows fully diminished seventh chords in the right hand in the third movement of Mozart's Piano Sonata No. 5 in the key of D major.
Function
The leading-tone seventh chord has a dominant function and may be used in place of V or V. Just as vii is sometimes considered an incomplete dominant seventh chord, a leading-tone seventh chord is often considered a "dominant ninth chord without root".)
For variety, leading-tone seventh chords are frequently substituted for dominant chords, with which they have three common tones: "The seventh chord founded upon the subtonic [in major]... is occasionally used. It resolves directly to the tonic... This chord may be employed without preparation."
Voice-leading
In contrast to leading-tone triads, leading-tone seventh chords appear in root position. The example below shows leading-tone seventh chords (in root position) functioning as dominants in a reduction of Mozart's Don Giovanni, K. 527, act 1, scene 13.
François-Joseph Fétis tunes the leading-tone seventh in major 5:6:7:9.
See also
Musica ficta
References
Sources
Further reading
Stainer, John, and William Alexander Barrett (eds.) (1876). A Dictionary of Musical Terms. London: Novello, Ewer and Co. New and revised edition, London: Novello & Co, 1898.
Diatonic functions
7.5 |
null | null | Commodore VIC-20 | eng_Latn | The Commodore VIC-20 (known as the VC-20 in Germany and the VIC-1001 in Japan) is an 8-bit home computer that was sold by Commodore Business Machines. The VIC-20 was announced in 1980, roughly three years after Commodore's first personal computer, the PET. The VIC-20 was the first computer of any description to sell one million units. It was described as "one of the first anti-spectatorial, non-esoteric computers by design...no longer relegated to hobbyist/enthusiasts or those with money, the computer Commodore developed was the computer of the future."
The VIC-20 was called VC-20 in Germany because the pronunciation of VIC with a German accent sounds like the German expletives "fick" or "wichsen". The term VC was marketed as though it were an abbreviation of VolksComputer ("people's computer," similar to Volkswagen and Volksempfänger).
History
Origin and marketing
The VIC-20 was intended to be more economical than the PET computer. It was equipped with 5 KB of static RAM and used the same MOS 6502 CPU as the PET. The VIC-20's video chip, the MOS Technology VIC, was a general-purpose color video chip designed by Al Charpentier in 1977 and intended for use in inexpensive display terminals and game consoles, but Commodore could not find a market for the chip.
As the Apple II gained momentum with the advent of VisiCalc in 1979, Jack Tramiel wanted a product that would compete in the same segment, to be presented at the January 1980 CES. For this reason Chuck Peddle and Bill Seiler started to design a computer named TOI (The Other Intellect). The TOI computer failed to materialize, mostly because it required an 80-column character display which in turn required the MOS Technology 6564 chip. However, the chip could not be used in the TOI since it required very expensive static RAM to operate fast enough.
In the meantime, new engineer Robert Yannes at MOS Technology (then a part of Commodore) designed a computer in his home dubbed the MicroPET and finished a prototype with help from Al Charpentier and Charles Winterble. With the TOI unfinished, when Jack Tramiel was shown the MicroPET prototype, he immediately said he wanted it to be finished and ordered it to be mass-produced following a limited demonstration at the CES.
As the new decade began, the price of computer hardware was dropping and Tramiel saw an emerging market for low-price computers that could be sold at retail stores to relative novices rather than professionals or people with an electronics or programming background. Radio Shack had been achieving considerable success with the TRS-80 Model I, a relatively low-cost machine that was widely sold to novices and in 1980 released the Color Computer, which was aimed at the home and educational markets, used ROM cartridges for software, and connected to a TV set.
The prototype produced by Yannes had very few of the features required for a real computer, so Robert Russell at Commodore headquarters had to coordinate and finish large parts of the design under the codename Vixen. The parts contributed by Russell included a port of the operating system (kernel and BASIC interpreter) taken from John Feagans design for the Commodore PET, a character set with the characteristic PETSCII, an Atari CX40 joystick-compatible interface, and a ROM cartridge port. The serial IEEE-488-derivative CBM-488 interface was designed by Glen Stark. It served several purposes, including costing substantially less than the IEEE-488 interface on the PET, using smaller cables and connectors that allowed for a more compact case design, and also complying with newly-imposed FCC regulations on RFI emissions by home electronics (the PET was certified as Class B office equipment which had less stringent RFI requirements). Some features, like the memory add-in board, were designed by Bill Seiler.
Altogether, the VIC 20 development team consisted of five people led by Michael Tomczyk the product manager, who recruited the group and dubbed them the VIC Commandos. Commodore founder Jack Tramiel initially gave Tomczyk the title VIC Czar and later appointed him product manager. Tomczyk insisted on several features including full-size typewriter keys, programmable function keys and a built-in RS-232 interface. Michael later contracted and co-designed a $100 modem, the VICModem, which became the first modem to sell 1 million units. According to one of the development team, Neil Harris, "[W]e couldn't get any cooperation from the rest of the company who thought we were jokers because we were working late, about an hour after everyone else had left the building. We'd swipe whatever equipment we needed to get our jobs done. There was no other way to get the work done! [...] they'd discover it was missing and they would just order more stuff from the warehouse, so everybody had what they needed to do their work."
At the time, Commodore had a glut of SRAM chips, so Tramiel decided these should be used in the new computer. The end result was arguably closer to the PET or TOI computers than to Yannes' prototype, albeit with a 22-column VIC chip instead of the custom chips designed for the more ambitious computers. As the amount of memory on the VIC-20's system board was very small even for 1981 standards, the design team could get away with using more expensive SRAM due to its lower power consumption, heat output, and less supporting circuitry. The original Revision A system board found in all silver-label VIC-20s used 2114 SRAMs and due to their tiny size (only 512 bytes per chip), ten of them were required to reach 5 KB of system RAM. The Revision B system board, found in rainbow logo VIC-20s (see below) switched to larger 2048-byte SRAMs which reduced the memory count to five chips: 2× 2048-byte chips + 3× 2114 (the 1024 × 4 bits) chips.
While newer PETs had the upgraded BASIC 4.0, which had disk commands and improved garbage collection, the VIC-20 reverted to the 8 KB BASIC 2.0 used on earlier PETs as part of another of the design team's goals, which was limiting the system ROMs to only 20 KB. Since Commodore's BASIC had been designed for the PET which had only limited audiovisual capabilities, there were no dedicated sound or graphics features, thus VIC-20 programmers had to use large numbers of POKE and PEEK statements for this. This was in contrast to the computer's main competitors, the Atari 400 and TRS-80 Color Computer, both of which had full-featured BASICs with support for the machines' sound and graphics hardware. Supplying a more limited BASIC in the VIC-20 would keep the price low and the user could purchase a BASIC extender separately if he desired sound or graphics commands.
While the TRS-80 Color Computer and Atari 400 had only RF video output, the VIC-20 instead had composite output, which provided a sharper, cleaner picture if a dedicated monitor was used. An external RF modulator was necessary to use the computer with a TV set, and had not been included internally so as to comply with FCC regulations (Commodore lobbied for and succeeded in getting them relaxed slightly by 1982, so the C64 had an RF modulator built in).
In April 1980, at a meeting of general managers outside London, Jack Tramiel declared he wanted a low-cost color computer. When most of the GMs argued against it, preferring Peddle's more sophisticated design, he said: "The Japanese are coming, so we must become the Japanese!" (in reference to the threats of low-cost systems from Japan). This was in keeping with Tramiel's philosophy which was to make "computers for the masses, not the classes". The concept was supported at the meeting by Tomczyk, newly hired marketing strategist and assistant to the president; Tony Tokai, General Manager of Commodore Japan, and Kit Spencer, the UK's top marketing executive. Peddle disagreed with the decision and left the company with other engineers, so an engineering team in Commodore Japan led by Yash Terakura helped finish the design. The VIC-20 was marketed in Japan as VIC-1001 before VIC-20 was introduced to the US.
When they returned to California from that meeting, Tomczyk wrote a 30-page memo detailing recommendations for the new computer, and presented it to Tramiel. Recommendations included programmable function keys (inspired by competing Japanese computers), full-size typewriter-style keys, and built-in RS-232. Tomczyk insisted on "user-friendliness" as the prime directive for the new computer, to engineer Terakura, and proposed a retail price of . He recruited a marketing team and a small group of computer enthusiasts, and worked closely with colleagues in the UK and Japan to create colorful packaging, user manuals, and the first wave of software programs (mostly games and home applications).
Scott Adams was contracted to provide a series of text adventure games. With help from a Commodore engineer who came to Longwood, Florida, to assist in the effort, five of Adams's Adventure International game series were ported to the VIC. They got around the limited memory of VIC-20 by having the games reside in a ROM cartridge instead of being loaded into main memory via cassette as they were on the TRS-80 and other machines. The first production run of the five cartridges generated over $1,500,000 in sales for Commodore.
While the PET was sold through authorized dealers the VIC-20 primarily sold at retail, especially discount and toy stores, where it could compete directly with game consoles. It was the first computer to be sold in K-Mart. Commodore took out advertisements featuring actor William Shatner (of Star Trek fame) as its spokesman, asking: "Why buy just a video game?" and describing it as "The Wonder Computer of the 1980s". Television personality Henry Morgan (best known as a panelist on the TV game show I've Got a Secret) became the commentator in a series of Commodore product ads.
The "20" in the computer's name was widely assumed to refer to the text width of the screen (although in fact the VIC-20 has 22-column text, not 20) or that it referred to the combined size of the system ROMs (8 KB BASIC+8 KB KERNAL+4 KB character ROM). Bob Yannes claimed that "20" meant nothing in particular and "We simply picked '20' because it seemed like a friendly number and the computer's marketing slogan was 'The Friendly Computer'. I felt it balanced things out a bit since 'Vic' sounded like the name of a truck driver."
In 1981, Tomczyk contracted with an outside engineering group to develop a direct-connect modem-on-a-cartridge (the VICModem), which at US$99 became the first modem priced under US$100. The VICModem was also the first modem to sell over 1 million units. VICModem was packaged with US$197.50 worth of free telecomputing services from The Source, CompuServe and Dow Jones. Tomczyk also created a SIG called the Commodore Information Network to enable users to exchange information and take some of the pressure off of Customer Support inquiries, which were straining Commodore's lean organization. In 1982, this network accounted for the largest traffic on CompuServe.
Evolution
The VIC-20 went through several variations in its three and a half years of production. First-year models (1980) had a PET-style keyboard with a blocky font while most VIC-20s made during 1981 had a slightly different keyboard also shared with early C64s. The rainbow logo VIC-20 was introduced in early 1983 and has the newer C64 keyboard with gray function keys and the Revision B motherboard. It has a similar power supply to the C64 PSU, although the amperage is slightly lower. A C64 "black brick" PSU is compatible with Revision B VIC-20s; however, the VIC's PSU is not recommended on a C64 if any external devices, such as cartridges or user port accessories, are installed, as it will overdraw the available power. Older Revision A VIC-20s cannot use a C64 PSU or vice versa as their power requirement is too high.
Decline
The VIC-20 was a bestselling computer, becoming the first personal computer to sell over a million. In total, 2.5 million computers were manufactured. In summer 1982, Commodore unveiled the a more advanced machine with of RAM and considerably improved sound and graphics. Initial sales of the C64 were slow, but took off in mid-1983. The VIC-20 was widely available for under $90 by that time. Commodore discontinued the VIC-20 in January 1985.
Perhaps the last new commercially available VIC-20 peripheral was the VIC-Talker, a speech synthesizer. Ahoy! wrote in January 1986, "Believe it or not, a new VIC accessory...We were as surprised as you."
Applications
The VIC-20's BASIC is compatible with the PET's, and the Datasette format is the same. Before the computer's release, a Commodore executive promised it would have "enough additional documentation to enable an experienced programmer/hobbyist to get inside and let his imagination work". Compute! favorably contrasted the company's encouragement of "cottage industry software developers" to Texas Instruments discouraging third-party software. Because of its small memory and low-resolution display compared to some other computers of the time, the VIC-20 was primarily used for educational software and games. However, productivity applications such as home finance programs, spreadsheets, and communication terminal programs were also made for the machine.
The VIC had a sizable library of public domain and freeware software. This software was distributed via online services such as CompuServe, BBSs, as well as offline by mail order and by user groups. Several computer magazines sold on newsstands, such as Compute!, Family Computing, RUN, Ahoy!, and the CBM-produced Commodore Power Play, offered programming tips and type-in programs for the VIC-20.
An estimated 300 commercial titles were available on cartridge and another 500+ were available on tape. A handful of disk applications were released.
The VIC's low cost led to it being used by the Fort Pierce, Florida, Utilities Authority to measure the input and output of two of their generators and display the results on monitors throughout the plant. The utility was able to purchase multiple VIC and C64 systems for the cost of one IBM PC compatible.
Technical specifications
The VIC-20 shipped with RAM, but of this is used for the video display and dynamic aspects of the ROM-resident Commodore BASIC and KERNAL (a low-level operating system). Only 3,583 bytes of BASIC program memory for code and variables are actually available on an unexpanded machine.
Ports and sockets
The VIC-20 has card edge connectors for program/expansion cartridges and a PET-standard Datassette tape drive. The VIC-20 did not originally have a disk drive; the VIC-1540 disk drive was released in 1981.
There is one Atari joystick port, compatible with the digital joysticks and paddles used with Atari VCS and Atari 8-bit family; a serial CBM-488 bus (a serial version of the PET's IEEE-488 bus) for daisy chaining disk drives and printers; a TTL-level "user port" with both RS-232 and Centronics signals (most frequently used as RS-232, for connecting a modem).
The VIC has a ROM cartridge port for games and other software as well as for adding memory to the machine. Port expander boxes from Commodore and other vendors allow more than one cartridge to be attached at a time. Cartridge size ranges from in size, although the latter was uncommon due to its cost.
The VIC-20 can be hooked into external electronic circuitry via joystick port, the "user port," or the memory expansion cartridge port, which exposes various analog to digital, memory bus, and other internal I/O circuits to the experimenter. PEEK and POKE commands from BASIC can be used to perform data acquisition from temperature sensors, control robotic stepper motors, etc.
Graphics
The graphics capabilities of the VIC chip (6560/6561) are limited but flexible. At startup the screen shows 176×184 pixels, with a fixed-color border to the edges of the screen. Since a PAL or NTSC screen has a 4:3 width-to-height ratio, each VIC pixel is much wider than it is high. The screen normally shows 22 columns and 23 rows of 8-by-8-pixel characters; it is possible to increase these dimensions up to 27 columns, but the characters would soon run out the sides of the monitor at about 25 columns. Just as on the PET, two different 256 character sets are included, the uppercase/graphics character set and the upper/lowercase set, and reverse video versions of both. Normally, the VIC-20 operates in high-resolution mode whereby each character is 8×8 pixels in size and uses one color. A lower-resolution multicolor mode can also be used with 4×8 characters and three colors each, but it is not used as often due to its extreme blockiness.
The VIC chip does not support a true bitmap mode, but programmers can define their own custom character sets. It is possible to get a fully addressable screen, although slightly smaller than normal, by filling the screen with a sequence of different double-height characters, then turning on the pixels selectively inside the RAM-based character definitions. The Super Expander cartridge adds BASIC commands supporting such a graphics mode using a resolution of 160×160 pixels. It is also possible to fill a larger area of the screen with addressable graphics using a more dynamic allocation scheme, if the contents are sparse or repetitive enough. This is used by the port of Omega Race.
The VIC chip has readable scan-line counters but cannot generate interrupts based on the scan position. The two VIA timer chips can serve this purpose through an elaborate programming technique, allowing graphics to be mixed with text above or below it, two different background and border colors, or more than 200 characters for the pseudo-high-resolution mode.
The VIC chip can process a light pen signal via the joystick port, but few appeared on the market.
The VIC chip outputs Luma+Sync and Chroma video signals, which is combined to create the VIC-20's composite video output. Commodore did not include an RF modulator inside the computer's case because of FCC regulations. It can either be attached to a dedicated monitor or a TV set using the external modulator included with the computer.
Sound
The VIC chip has three pulse wave generators and a white noise generator with an overall volume control and mono output. Each pulse wave generator has a range of three octaves located on the scale about an octave apart, giving a total range of about five octaves.
Memory expansion
The VIC-20's RAM is expandable through the cartridge port via a RAM cartridge. RAM cartridges were available in several sizes: (with or without an included "Super Expander" BASIC extension ROM), , , and , the latter two only from third-party vendors. The internal memory map is dramatically reorganized with the addition of each size cartridge, leading to a situation where some programs only work if the right amount of memory is present (to cater to this, the cartridges had switches, and the cartridges had software setups, allowing the RAM to be enabled in user-selectable memory blocks). Since the VIC-20 was designed to use SRAM rather than DRAM, the system board has no provisions for DRAM refresh.
The memory mapping of the VIC-20 can vary depending on system configuration. With no expanders installed, free user memory starts at $1000 and extended up to $1DFF, with the video buffer placed at $1E00-$1FFF. Below $1000 was a gap from $400-$FFF which could be filled with of expansion RAM, which if installed would place the BASIC program area at $400. If an or larger expander was used, screen memory began at $1000 and BASIC program text at $1200.
The VIC-20 allows two locations for color RAM, either at $9600 or $9400. The default for unexpanded machines is at $9600, and with an or larger expander, the color RAM would be moved to $9400; however, the user can freely set it to either location via the register at $9002.
The normal location for ROM cartridges is at $A000–$BFFF. On power up, the kernel ROM checks for an ID header and if found jumps to the specified starting address. Larger cartridges use the second half of ROM either at $2000 or $6000. A few cartridges, including Scott Adams adventures, load entirely in the $2000-$7FFF area. Since the kernel can only autostart ROMs located at $A000, such programs has to be manually launched from BASIC via the SYS command.
Commodore's official RAM expansion cartridges were only available up to a maximum of worth of additional memory, but third party cartridges can provide up to and sometimes included DIP switches to map the additional RAM to user-selectable address space.
Unlike the PET, the VIC-20 does not include a machine language monitor, but Commodore offered them on disk, tape, or cartridge, with several different executables to load into various memory locations. The monitor programs were the same as the PET monitor, but added a mini-assembler instead of requiring the user to enter hexadecimal opcodes.
The cartridges allowed adding up to to the BASIC user memory; together with the built-in user memory, this gave a maximum of for BASIC programs and variables. The extra could usually be used in one of two ways, set by switches:
Either it could be mapped into the address space reserved for ROM cartridges, which sat "behind" the I/O register space and thus was not contiguous with the rest of the RAM. This allowed running many cartridge-based games from disk or tape and was thus very useful for software pirates; especially if the RAM expansion allowed switching off writing to its memory after the game was loaded, so that the memory behaved exactly like ROM.
Or, of the could be mapped into the same memory "hole" that the cartridge used, letting lie fallow. These were contiguous with the rest of RAM, but couldn't be used to expand BASIC space to more than , because the display data would have had to be moved to cartridge RAM, which was not possible.
Reception
Describing it as "an astounding machine for the price", Compute! in 1981 expected the VIC-20 would be popular in classrooms and homes with small children, with "excellent graphic and sound capabilities". While predicting the 22-column screen was "too small to support any but the most rudimentary business applications" the magazine observed that "at a price of $299, that is hardly the point", stating that "the VIC will provide very stiff competition to the TRS-80 Color Computer" and "is a much more valuable computer literacy tool than" other products like the TRS-80 Pocket Computer. Compute! concluded "VIC will create its own market, and it will be a big one". While also noting the small screen size and RAM, BYTE stated that the VIC 20 was "unexcelled as low-cost, consumer-oriented computer. Even with some of its limitations...it makes an impressive showing against...the Apple II, the Radio Shack TRS-80, and the Atari 800". The magazine praised the price ("Looking at a picture...might cause you to think $600 would be a fair price...But it does not cost $600—the VIC 20 retails for $299.95"), keyboard ("the equal of any personal-computer keyboard in both appearance and performance. This is a remarkable achievement, almost unbelievable considering the price of the entire unit"), graphics, documentation, and ease of software development with the KERNAL.
See also
VICE, a VIC-20 emulator
List of Commodore VIC-20 games
Footnotes
References
Further reading
BYTE in 1983 published a series of technical articles about the VIC-20:
External links
OLD-COMPUTERS.COM online-museum VIC-20 page
VIC-20 Programmers reference guide and more
6502-based home computers
Computer-related introductions in 1980 |
null | null | Prince William, Duke of Cambridge | eng_Latn | Prince William, Duke of Cambridge, (William Arthur Philip Louis; born 21 June 1982) is a member of the British royal family. He is the elder son of Charles, Prince of Wales, and Diana, Princess of Wales. Since birth, he has been second in the line of succession to the British throne.
Born in St Mary's Hospital, London, William was educated at Wetherby School, Ludgrove School and Eton College. He spent parts of his gap year in Belize and Chile before earning a Scottish Master of Arts degree in geography at the University of St Andrews. William then trained at Royal Military Academy Sandhurst prior to serving with the Blues and Royals. In April 2008, William graduated from Royal Air Force College Cranwell, joining RAF Search and Rescue Force in early 2009. He served as a full-time pilot with the East Anglian Air Ambulance from July 2015 for two years.
The Duke performs official duties and engagements on behalf of the Queen. He holds patronage with over 30 charitable and military organisations, including the Tusk Trust, Centrepoint, and London's Air Ambulance Charity. He undertakes projects through The Royal Foundation, with his charity work revolving around mental health, conservation, and emergency workers. In December 2014, he founded the "United for Wildlife" initiative, which aims to reduce worldwide illegal wildlife trade. In April 2016, the Cambridges and Prince Harry initiated the mental health awareness campaign "Heads Together" to encourage people to open up about their mental health issues. In October 2020, William launched the Earthshot Prize, a £50 million initiative to incentivise environmental solutions over the next decade.
In 2011, William was made Duke of Cambridge preceding his marriage to Catherine Middleton. The couple have three children: Prince George, Princess Charlotte, and Prince Louis of Cambridge.
Early life
Prince William was born at St Mary's Hospital, London on 21 June 1982 as the first child of Charles, Prince of Wales (heir apparent to Queen Elizabeth II) and Diana, Princess of Wales. His names, William Arthur Philip Louis, were announced by Buckingham Palace on 28 June. He was baptised by the Archbishop of Canterbury, Robert Runcie, in the Music Room of Buckingham Palace on 4 August, the 82nd birthday of his paternal great-grandmother Queen Elizabeth The Queen Mother. William had six godparents: former King Constantine II of Greece (his paternal second cousin once removed); Princess Alexandra, The Honourable Mrs Ogilvy (his paternal first cousin twice removed); the Duchess of Westminster; Lady Susan Hussey; Lord Romsey (his paternal second cousin once removed); and Sir Laurens van der Post.
He was the first child born to a Prince and Princess of Wales since Prince John in 1905. William accompanied his parents on their 1983 tour of Australia and New Zealand, when he was nine months old, as his first trip overseas. He travelled with his family to Canada in 1991 and 1998.
Prince William and his younger brother, Harry, were raised at Kensington Palace in London, and Highgrove House in Gloucestershire. Known informally as "Wills" within the family, William was nicknamed "Wombat" by his mother, who wished him and his brother to obtain broader life experiences than those usually available to royal children. She took them to Walt Disney World and McDonald's, AIDS clinics, shelters for the homeless, and bought them items typically owned by teenagers, such as video games. His parents divorced in 1996. Diana died in a car accident in the early hours of 31 August 1997. William, then aged 15, together with his 12-year-old brother and their father, were staying at Balmoral Castle at the time. The Prince of Wales waited until his sons awoke the following morning to tell them about their mother's death. William accompanied his father, brother, paternal grandfather Prince Philip, Duke of Edinburgh, and his maternal uncle Charles Spencer, 9th Earl Spencer, at his mother's funeral. William and Harry walked behind the funeral cortège from Kensington Palace to Westminster Abbey.
Education
William was educated at independent schools, starting at Jane Mynors' nursery school and the pre-preparatory Wetherby School, both in London. Following this, he attended Ludgrove School near Wokingham, Berkshire, and was privately tutored during summers by Rory Stewart. At Ludgrove, he participated in football, swimming, basketball, clay pigeon shooting, and cross country running. He sat the entrance exam to Eton College and was admitted. There, he studied Geography, Biology, and History of Art at A-Level, obtaining an 'A' in Geography, a 'C' in Biology, and a 'B' in History of Art. At Eton, he took up water polo and continued to play football, captaining his house team.
The decision to place William in Eton went against the family tradition of sending royal children to Gordonstoun, which his grandfather, father, two uncles, and two cousins all attended. Diana's father and brother both attended Eton. The royal family and the tabloid press agreed William would be allowed to study free from intrusion in exchange for regular updates about his life. John Wakeham, chairman of the Press Complaints Commission, said of the arrangement, "Prince William is not an institution; nor a soap star; nor a football hero. He is a boy: in the next few years, perhaps the most important and sometimes painful part of his life, he will grow up and become a man."
After completing his studies at Eton, William took a gap year, during which he took part in British Army training exercises in Belize, worked on English dairy farms, visited Africa, and for ten weeks taught children in southern Chile. As part of the Raleigh International programme in the town of Tortel, William lived with other young volunteers, sharing in the common household chores—including cleaning the toilet—and also volunteered as a guest disc jockey at a local radio station. His interest in African culture prompted him to teach himself Swahili.
By 2001, William was back in the United Kingdom and had enrolled at the University of St Andrews. The extra attention did not deter him; he embarked on a degree course in Art History, later changing his main subject to Geography. William wrote his dissertation on the coral reefs of Rodrigues in the Indian Ocean and graduated with Scottish Master of Arts degree with upper second class honours in 2005. While at university, he represented the Scottish national universities water polo team at the Celtic Nations tournament in 2004. He was known as "Steve" by other students to avoid any journalists overhearing and realising his identity. William returned to St Andrews alongside his wife in February 2011 as patron of the university's 600th Anniversary Appeal.
Upon graduation from university, William interned in land management at Chatsworth House and in banking at HSBC. To prepare for his eventual management of the Duchy of Cornwall, in 2014, he enrolled in a vocational agricultural management course at Cambridge, which was organised by the Cambridge Programme for Sustainability Leadership (CPSL), of which his father is patron. According to a CNN report in 2014, the duchy is "a £760 million (about US$1.25 billion) entity established in 1337 to provide a private income for use by the reigning monarch's eldest son", which William will inherit when his father becomes king.
Military and air ambulance service
Military training and secondments
Having decided to follow a military career, he was admitted to the Royal Military Academy Sandhurst in January 2006. William officially received his commission as a lieutenant at midnight. As "Lieutenant Wales"—a name based on his father's title Prince of Wales—he followed his younger brother into the Blues and Royals as a troop commander in an armoured reconnaissance unit, after which he spent five months training for the post at Bovington Camp, Dorset.
William's position as second-in-line to the throne and the convention of ministers advising against placing that person into dangerous situations cast doubts on his chances of seeing combat, which increased after Prince Harry's deployment was cancelled in 2007 due to "specific threats". William, instead, went on to train in the Royal Navy and Royal Air Force, obtaining his commission as a sub-lieutenant in the former and flying officer in the latter—both broadly equivalent to the army rank of lieutenant.
After completing his training, William undertook an attachment with the Royal Air Force at RAF Cranwell. Upon completing the course he was presented with his RAF wings by his father, who had received his own wings after training at Cranwell. During this secondment, William flew to Afghanistan in a C-17 Globemaster that repatriated the body of Trooper Robert Pearson. William was then seconded to train with the Royal Navy. He completed an accelerated Naval Officer training course at the Britannia Royal Naval College. Whilst serving on HMS Iron Duke in June 2008, William participated in a £40m drug bust in the Atlantic, north-east of Barbados. He was a part of the crew on the Lynx helicopter which helped seize 900 kg of cocaine from a speedboat.
Royal Air Force service
In January 2009, William transferred his commission to the RAF and was promoted to Flight Lieutenant. He trained to become a helicopter pilot with the RAF's Search and Rescue Force. In January 2010, he graduated from the Defence Helicopter Flying School at RAF Shawbury. On 26 January 2010, he transferred to the Search and Rescue Training Unit at RAF Valley, Anglesey, to receive training on the Sea King search and rescue helicopter; he graduated in September 2010. This made him the first member of the British royal family since Henry VII to live in Wales.
William's first rescue mission as co-pilot of an RAF Sea King was a response to an emergency call from Liverpool Coastguard on 2 October 2010. In November 2011, he participated in a search-and-rescue mission involving a cargo ship that was sinking in the Irish Sea; William, as a co-pilot, helped rescue two sailors.
William was deployed to the Falkland Islands for a six-week tour with No. 1564 Flight from February to March 2012. The Argentine government condemned the Duke's deployment to the islands close to the 30th anniversary of the beginning of the Falklands War as a "provocative act". In June 2012, Prince William gained a qualification to be captain or pilot in command of a Sea King rather than a co-pilot. His active service as an RAF search-and-rescue pilot ended in September 2013. He later became patron to the Battle of Britain Memorial Flight.
Air ambulance pilot
In 2014, it was announced that William would accept a full-time role as a pilot with the East Anglian Air Ambulance (EAAA) based at Cambridge Airport. Despite his qualifications as a military helicopter pilot, William needed a civil pilot's licence and further training before being permitted to take command of the Air Ambulance. Although his position was paid, Kensington Palace announced that William would donate his full salary to the EAAA charity. He underwent part of his training as an EAAA pilot at Norwich Airport. On 13 July 2015, William started his new job, which he felt was a natural progression from his previous role as an RAF search-and-rescue pilot. The Duke described working irregular shifts and dealing mostly with critical care cases. He also publicly discussed the consequences, witnessing intensive trauma and bereavement as an emergency worker, stating that it impacted his mental health and personal life. BBC has written that the Duke was "exposed to the National Health Service in a way that no other senior royal has been or possibly ever will be."
William left his position with EAAA in July 2017 to assume full-time royal duties on behalf of his grandmother. After supporting an anniversary campaign for London's Air Ambulance Charity in 2019, the Duke became the charity's official patron in March 2020. In May 2020, he granted permission to the charity to use Kensington Palace's private lawn to refuel during the COVID-19 pandemic. To mark Air Ambulance Week 2020, he wrote a letter thanking air ambulance workers, stating his "profound respect" for the community, particularly during the "immeasurably difficult" outbreak, and stated that "the country owes you an enormous debt of gratitude."
Personal life
Bachelorhood
In 2001, William met Catherine Middleton while they were students in residence at St Salvator's Hall at the University of St Andrews. She reportedly caught William's attention at a charity fashion show on campus. The couple began dating in 2003. During their second year, William shared a flat with Middleton and two other friends. From 2003 to 2005, they both resided at Balgove House on the Strathtyrum estate with two roommates.
Their relationship was followed so closely by the tabloid press that bookmakers took bets on the possibility of marriage, and the retail chain Woolworths produced memorabilia bearing their likenesses. Media attention became so intense that William formally asked the press to keep their distance from Middleton. On 15 December 2006, Middleton attended Prince William's Passing Out Parade at the Royal Military Academy Sandhurst.
In April 2007, William and Middleton ended their relationship. Middleton and her family attended the Concert for Diana in July 2007 at Wembley Stadium, where she and Prince William sat two rows apart. The couple were subsequently seen together in public on a number of occasions and news sources stated that they had "rekindled their relationship". Middleton was in attendance during the Order of the Garter procession ceremony at Windsor Castle in June 2008, where Prince William was made a Royal Knight of the Garter. In June 2010, the couple moved into a cottage on the Bodorgan Estate in Anglesey, Wales, where William resided during his RAF search-and-rescue training and subsequent career.
Marriage and children
On 16 November 2010, Clarence House announced that William and Catherine were to marry; the couple had become engaged in Kenya in October. The engagement ring given by William to Catherine had belonged to his mother. The wedding took place on 29 April 2011 in Westminster Abbey, London. A few hours before the ceremony, William's new titles Duke of Cambridge, Earl of Strathearn, and Baron Carrickfergus were announced. Estimates of the global audience for the wedding ranged around 300 million or more, whilst 26 million watched the event live in Britain alone. The couple were given the country home, Anmer Hall, on the Sandringham Estate, as a wedding gift from the Queen. The Duke and Duchess owned an English Cocker Spaniel, Lupo, from December 2011 to November 2020.
Catherine's first pregnancy was announced on 3 December 2012. She was admitted on 22 July 2013 to the Lindo Wing of St Mary's Hospital, London, where Prince William had been delivered. Later that day, she gave birth to Prince George. On 8 September 2014, it was announced that the Duchess of Cambridge was pregnant with her second child. She was admitted on 2 May 2015 to the same hospital and gave birth to Princess Charlotte. The Duchess's third pregnancy was announced on 4 September 2017; Prince Louis was born on 23 April 2018. The family officially reside at Kensington Palace.
William is the godfather of Prince Constantine Alexios of Greece and Denmark (b. 1998), a distant relation though his grandfather, the Duke of Edinburgh, and Mia Tindall (b. 2014), the eldest child of his paternal cousin, Zara Tindall.
Wealth and inheritance
William and his brother Harry inherited the "bulk" of the £12.9 million left by their mother on their respective 30th birthdays, a figure that had grown since her 1997 death to £10 million each in 2014. In 2002 The Times reported that William would also share with his brother a payment of £4.9 million from trust funds established by their great-grandmother, Queen Elizabeth The Queen Mother, on their respective 21st birthdays and would share a payment of £8 million upon their respective 40th birthdays. As the eldest son of the heir-apparent, William is expected to inherit the Duchy of Cornwall, which would bring him an additional income.
In 2014, the brothers inherited their mother's wedding dress along with many other of her personal possessions including dresses, diamond tiaras, jewels, letters, and paintings. The brothers also received the original lyrics and score of "Candle in the Wind" by Bernie Taupin and Elton John as performed by John at Diana's funeral.
Health
On 3 June 1991, William was admitted to Royal Berkshire Hospital after being accidentally hit on the forehead by a fellow pupil wielding a golf club. He suffered a depressed fracture of the skull and was operated on at Great Ormond Street Hospital, resulting in a permanent scar. In a 2009 interview, he dubbed this scar a "Harry Potter scar" and said, "I call it that because it glows sometimes and some people notice it—other times they don't notice it at all".
On 1 November 2020, it was reported that William had tested positive for COVID-19 in April but decided not to alert the media to 'avoid alarming the nation'. The Daily Telegraph reported he had been "very ill" and had isolated away from his family; other sources say that he had not been seriously ill, not bed-ridden and working for most of the time.
Public life
At the age of 21, William was appointed a Counsellor of State; he first served in that capacity when the Queen attended the Commonwealth Heads of Government Meeting in 2003. On his 21st birthday, William accompanied his father on a visit to Wales, visiting the Anglesey Food Festival and opening a centre for the homeless in Newport. Upon graduating from university, William began royal duties in support of, and on behalf of, the Queen at official events, public engagements, and overseas tours. In July 2005, he embarked on his first solo overseas tour, travelling to New Zealand to participate in World War II commemorations. For the 30th anniversary of his father's charity The Prince's Trust, William and his brother were jointly interviewed for the first time by television personalities Ant & Dec. According to author Tina Brown, he had, like his father, expressed a desire to become Governor-General of Australia. Prime Minister of Australia John Howard expressed his wish for the position to be held by an Australian citizen. In 2009, the Queen set up a private office for William with David Manning as his adviser. Manning accompanied him in January 2010 as he toured Auckland and Wellington; William opened the new building of the Supreme Court of New Zealand and was welcomed by a Māori chief.
In June 2010, William and his brother visited Botswana, Lesotho, and South Africa, visiting projects relating to wildlife, sport, and young children. In November 2010, he attended a memorial service held on Remembrance Day at Camp Bastion, Afghanistan. In March 2011, William visited Christchurch, New Zealand, shortly after the earthquake, and spoke at the memorial service at Hagley Park on behalf of his grandmother. He also travelled to Australia to visit areas affected by flooding in Queensland and Victoria. In April 2011, Time magazine selected him as one of the 100 Most Influential People in the World. In May 2011, the Duke and Duchess of Cambridge met with U.S President Barack Obama and First Lady Michelle Obama at Buckingham Palace. The couple toured Canada in summer 2011, attending Canada Day celebrations on Parliament Hill. On 2 November, the Duke and Duchess visited the UNICEF Supply Division for malnourished children in Copenhagen, Denmark.
William and Catherine served as ambassadors for the 2012 Summer Olympics in London, during multiple sporting events throughout the games. In September 2012, they toured Singapore, Malaysia, Tuvalu, and the Solomon Islands as part of the Queen's Diamond Jubilee celebrations. The Duke and Duchess attended further commemorations of the Jubilee throughout the year, including the Thames Diamond Jubilee Pageant in July. The Duke hosted his first investiture ceremony at Buckingham Palace in October 2013. In April 2014, William and Catherine undertook a royal tour to New Zealand and Australia with their son, Prince George. The itinerary included visiting the Plunket Society for children and visiting fire-damaged areas in New South Wales. In June 2014, the couple visited France to attend the commemoration of the 70th anniversary of the Normandy landings at Gold Beach. In September 2014, the Duke visited Malta to commemorate its 50th independence anniversary, substituting for his wife after the announcement of her second pregnancy. On 21 October, the Duke and Duchess met the President of Singapore, Tony Tan, during his state visit to the UK. In December 2014, the William met with President Obama in the Oval Office, and made a speech at the World Bank in Washington, D. C., condemning the illegal trade in wildlife. In December 2014, the couple visited New York and attended a charity dinner at the Metropolitan Museum of Art.
In February 2015, the Duke visited Japan, meeting with Emperor Akihito and Empress Michiko at the Imperial Palace and visiting survivors devastated by the 2011 tsunami. From 1 to 4 March, the Duke visited the Chinese cities Beijing, Shanghai, and Yunnan and met with President Xi Jinping. It was the first royal visit to mainland China in almost three decades. In April 2016, the William and his wife undertook a tour to India and Bhutan. Activities included visiting children's charities such as Childline India, as well as a visit to Lingkana Palace. Later that month, the couple met again with the Obamas at Kensington Palace. In April 2016, William and Catherine toured to India and Bhutan. The couple toured Canada once again in September 2016. In November 2016, he visited Vietnam, meeting with Prime Minister Nguyen Xuan Phuc and touring local primary schools. Countries visited by the couple in 2017 include France, Poland, Germany, and Belgium. In January 2018, the couple visited Sweden and Norway. The visits, which were, like others, requested by the Foreign Office, were interpreted to benefit UK-European relations post Brexit. In June 2018, the Duke toured Jordan, Israel and Palestine.
In February 2019, William and Catherine carried out a two-day visit of Northern Ireland, visiting Belfast, Fermanagh, and Ballymena. In March and April 2019, William spent three weeks working for MI6, MI5 and GCHQ. The Duke and Duchess toured Pakistan in October 2019, which was the royal family's first visit to the country in 13 years. In December 2019, William visited Kuwait and Oman, commemorating the 120th anniversary of the Anglo-Kuwaiti Agreement of 1899. In March 2020, the couple carried out a three-day tour of Ireland, visiting County Meath, Kildare, and Galway. In October 2020, the Duke and Duchess met Volodymyr Zelensky, the President of Ukraine, and First Lady Olena Zelenska, at Buckingham Palace, the first royal engagement held at the residence since the start of the COVID-19 pandemic. In December, the couple embarked on a three-day tour of England, Scotland, and Wales via the British royal train "to pay tribute to the inspiring work of individuals, organisations and initiatives across the country" in 2020. Prime Minister Boris Johnson expressed his support for the initiative, while First Minister of Scotland Nicola Sturgeon criticised the tour, citing travel restrictions; UK, Scottish and Welsh governments were consulted before planning the tour. In William's capacity as Lord High Commissioner to the General Assembly of the Church of Scotland, the couple toured Edinburgh, Fife and Orkney in May 2021. In Cornwall on 11 June 2021, William and Catherine attended the G7 summit for the first time. They also attended a reception, where the Duke and his father discussed governmental and corporate solutions to environmental problems.
Charity work and interests
Humanitarian and environmental patronages
William became aware of HIV/AIDS in the mid-1990s when he accompanied his mother and brother on visits to shelters and clinics for patients. In January 2005, William and his brother volunteered at a British Red Cross aid distribution centre to pack emergency supplies for countries affected by the 2004 Boxing Day tsunami. In September that year, William granted his patronage to Centrepoint, a charity that assists the homeless. In December 2009, he, as part of a Centrepoint-organised event, spent the night sleeping bag near the Blackfriars Bridge to raise awareness of the experiences of homeless youth. The Duke opened their new facility, Apprenticeship House, in November 2019 to mark their 50-year anniversary. In 2005, William worked in the children's unit at The Royal Marsden Hospital, his mother's former patronage, for two days of work experience; he also assisted in the medical research, catering, and fundraising departments. In May that year, he spent two weeks in North Wales with Mountain Rescue England and Wales. In May 2007, William became patron of both organisations. In October 2020, the Duke laid the foundation stone of the hospital's Oak Cancer Centre, 30 years after his mother did the same for their Chelsea Wing in 1990.
Prince William became a patron of the Tusk Trust in December 2005, a charity that works towards conserving wildlife and initiating community development, including providing education, across Africa. He became associated with the organisation after he witnessed its work first hand in Africa. Stating that "rural African initiatives that foster education, responsibility and participation in the local community light the way to conservation", he carried out his first official duty with the trust in launching a bike ride across the African continent in 2008. Later, William helped with launching the Tusk Conservation Awards, which have been presented to selected environmental activists annually since 2013. William has occasionally commented on the effects of overpopulation on the wildlife of Africa, but his remarks have been criticised for not taking resource consumption and population density into consideration, both of which affect wildlife in rich and developed countries.
In 2007, William and Harry organised the Concert for Diana, in memory of their mother, which benefitted the charities and patronages of Diana, William, and Harry. In 2010, he also became a patron of 100 Women in Hedge Funds Philanthropic Initiatives. William succeeded Lord Attenborough in 2010 as the fifth president of the British Academy of Film and Television Arts. In March 2011, the Duke and Duchess set up a gift fund held by The Foundation of Prince William and Prince Harry to allow well-wishers who wanted to give them a wedding gift to donate money to charities instead. The gift fund supported 26 charities of the couple's choice, incorporating the armed forces, children, the elderly, art, sport and conservation. The charity has since been renamed The Royal Foundation of the Duke and Duchess of Cambridge.
In December 2014, the Duke founded the United for Wildlife Transport Taskforce, which aims to reduce worldwide illegal wildlife trade. The Duke has spoken out for LGBT rights as part of his work against cyberbullying, stating the importance of being "proud of the person you are" and discussing the effects of online abuse and discrimination. In 2016, he appeared in the July issue of Attitude and became the first member of the royal family to be featured on the cover of a gay magazine. He was recognised at the British LGBT Awards in May 2017. In 2016, the Royal Foundation launched multiple mental health initiatives, including Heads Together, a campaign led by the Duke and Duchess of Cambridge and Prince Harry to de-stigmatise mental health. Legacy programmes include Mental Health at Work, launched in September 2018 to change the approach to workplace mental health in the UK, as well as Heads Up, launched in May 2019 in partnership with the Football Association, utilising football to affect the conversation surrounding mental health in adults. Later that month, the Duke and Duchess together with William's brother and sister-in-law launched Shout, the UK's first 24/7 text messaging service for those who suffer from mental issues. In October 2019, the Duke of Cambridge, together with other members of the royal family, voiced a PSA video for Public Health England "as part of its Every Mind Matters program". William has cited his interest in mental health to his experiences as an air ambulance pilot, as well as his work with homelessness, veterans welfare, and his wife's advocacy on addiction.
William has been patron of homelessness charity The Passage since 2019 after first visiting the center in 1992 with his mother. In October 2020, he wrote the introduction to the organisation's 40th-anniversary fundraising cookbook, discussing the importance of helping victims of homelessness during the COVID-19 pandemic. In December 2020, the Duke volunteered at the charity to help prepare donation bags for homeless residents in emergency hotel accommodations and spoke with residents about their experiences. In March 2020, the Duke appeared in a video for the National Emergencies Trust, launching a fundraising appeal to help charities during the pandemic. The appeal raised £11 million in its first week, eventually totalling to £90 million, with the money going out to "front line charities" and to the UK Community Foundations to be distributed among "local community foundations". In April 2020, he officially became the patron of the organisation. In late March 2020, the Duke and Duchess of Cambridge began supporting a new mental health initiative by the Public Health England amidst the coronavirus pandemic. He made a surprise appearance in The Big Night In, a 20 April 2020 telethon held during the COVID-19 pandemic, in a skit which he held a video call with Stephen Fry, who revised his role as (a descendant of) Lord Melchett, from the Blackadder series. Later that month, the Duke and Duchess announced Our Frontline, an initiative providing mental health support to emergency medical workers.
In May 2020, the Duke and Duchess of Cambridge's recorded radio message for Mental Health Awareness Week was broadcast across all the stations in the UK. In May and June 2020, the couple, alongside their children, delivered food parcels made on the Sandringham Estate to local isolated pensioners during the COVID-19 pandemic. In June 2020, the Duke of Cambridge stated that he had been serving as a volunteer on the Shout hotline during the pandemic. In September 2020, the Duke established the Emergency Responders Senior Leaders Board, commissioned by the foundation to research the mental health and wellbeing of emergency responders. The project is in partnership with King's College London and the Open University. After two years of research, the Duke launched the Earthshot Prize in October 2020, designed to provide funding and incentive for environmental solutions over the next decade. The Prize is slated to be given every year from 2021 until 2030 to five winners each year, in accordance with five categories detailing the restoration of nature, air cleanliness, ocean conservation, waste-free living, and climate change. The selection process will be performed by the Duke, alongside a council of judges from six continents, overseen by a panel of experts. The first awards ceremony is slated to take place in London in autumn 2021. Following the launch, William gave a TED Talk on environmental protection and conservation as part of the TED Countdown climate change initiative. Later that month, the Duke took over the patronages of Flora and Fauna International and the British Trust for Ornithology, passed on from the Queen and Prince Philip. In the same month, he appeared in an ITV Documentary titled Prince William: A Planet For Us All to discuss environmental issues. In December 2020, the Duke and Duchess became joint patrons of NHS Charities Together. In February 2021, William visited a vaccination centre in King's Lynn and later encouraged use of the vaccine, denouncing false information that could cause vaccine hesitancy. In May 2021, he got his first dose of COVID-19 vaccine by NHS staff at the Science Museum in London. In September 2021, it was reported that William had helped an Afghan officer who was a graduate of the Royal Military Academy Sandhurst and an assistant to the British troops be evacuated from the Kabul airport along with more than 10 members of his family amid the 2021 Taliban offensive.
Sport
William often plays polo to raise money for charity. He is a fan of football, and supports the English club Aston Villa. He became President of England's Football Association in May 2006 and vice-royal patron of the Welsh Rugby Union (WRU) in February 2007, supporting the Queen as patron. The same year, the WRU's decision to name a new cup for test matches between Wales and South Africa the Prince William Cup caused controversy; some believed it would have been more appropriate to name it after Ray Gravell. In December 2010, William and Prime Minister David Cameron attended a meeting with FIFA vice-president Chung Mong-joon at which Chung suggested a vote-trading deal for the right to host the 2018 World Cup in England. The English delegation reported the suggestion to FIFA's ethics investigator because they considered vote-swapping to be a violation of anti-collusion rules. In 2011, William as President of the English FA, voted against Australia's 2022 FIFA bid and instead voted for South Korea; despite being the country's future heir. In 2020, again as President of the English FA, he voted against the joint Australia–New Zealand 2023 FIFA Women's World Cup bid and instead voted for Colombia.
In February 2021, following an investigation into racism directed at Marcus Rashford, the Duke released a statement as president of the FA, denouncing the "racist abuse... whether on the pitch, in the stands, or on social media" as "despicable" and stating that "we all have a responsibility" to create an environment of tolerance and accountability. In April 2021, William criticised the planned breakaway competition The Super League, adding that he "share[d] the concerns of fans about the proposed Super League and the damage it risks causing to the game we love." In July 2021, he condemned racist attacks against England football players following their loss at the UEFA Euro 2020 finale.
In 2006, William, along with other Sandhurst officers, took part in a run to support the charity Sport Relief, as he had done in 2004 with a team from Clarence House. In May 2007, William became patron of the English Schools' Swimming Association. In 2012, together with the Duchess of Cambridge and Prince Harry, William launched Coach Core. The program was set up following the 2012 Olympics and provides apprenticeship opportunities for people who desire to pursue a career as a professional coach. In 2013, he succeeded his grandfather Prince Philip, Duke of Edinburgh, as president of the UK charity Fields in Trust. In 2014 he and the Duchess were awarded Honorary Life Membership of the Marylebone Cricket Club. In May 2020, the Duke of Cambridge appeared in a BBC One Documentary titled Football, Prince William and Our Mental Health as a part of a campaign to promote men to discuss their mental issues using football as a common medium.
Both William and his brother are enthusiastic motorcyclists; William owns a Ducati 1198 S Corse. In May 2014, William, like his father and paternal grandfather, became president of the British Sub-Aqua Club (BSAC). He enthusiastically took part in a bandy event in Stockholm in January 2018.
Privacy and the media
The death of Diana, Princess of Wales, in Paris while being chased by paparazzi in 1997, influenced the Duke's attitude towards the media. The Duke and his wife have asked that, when off-duty, their privacy should be respected.
On 13 November 2005, an article appeared in the News of the World written by royal editor Clive Goodman, that claimed that Prince William was in the process of borrowing a portable editing suite from ITV royal correspondent Tom Bradby. Prince William noted that another equally improbable leak had recently taken place regarding an appointment he had made with a knee surgeon. After some discussion, the Prince and Bradby concluded it was likely that their voicemails were being accessed. An investigation under Deputy Assistant Commissioner Peter Clarke concluded that the compromised voice mail accounts belonged to Prince William's aides, including Jamie Lowther-Pinkerton, and not the Prince himself. However, Clive Goodman later stated that he had hacked William's phone on 35 occasions.
In September 2012, the French edition of Closer and Italian gossip magazine Chi published photographs of the Duchess of Cambridge sun-bathing topless while on holiday at the Château d'Autet (a private château on a 260-ha estate 71 km north of Aix-en-Provence). Analysts from The Times believed the photographs were taken from the D22 (Vaucluse) road half a kilometre from the pool—a distance that would require an 800-mm or a 1000-mm lens. On 17 September 2012, the Duke and Duchess filed a criminal complaint with the French prosecution department and launched a claim for civil damages at the Tribunal de Grande Instance de Nanterre. The following day the courts granted an injunction against Closer prohibiting further publication of the photographs and announced a criminal investigation would be initiated. Under French law, punitive damages cannot be awarded but intrusions of privacy are a criminal offence carrying a maximum jail sentence of one year and a fine of up to €45,000 for individuals and €225,000 for companies. In September 2017, Closer was fined €100,000 and its editor Laurence Pieau and owner Ernesto Mauri were each fined €45,000.
In October 2014, William and Catherine sent a legal letter to a freelance photographer who had put their son George and his nanny "under surveillance", asking the individual to stop "harassing and following" them. In August 2015, Kensington Palace published a letter detailing what it stated were the "dangerous" and invasive efforts of the media to get paparazzi pictures of Prince George and Princess Charlotte. Jason Knauf, communications secretary to the Cambridges, wrote the letter to media standards organisations in various countries.
In March 2017, a video of William "dad dancing" at a nightclub in Verbier, Switzerland surfaced in tabloid media. At the time, he was on a skiing holiday, and was reportedly observed partying alongside friends and models. The incident garnered commentary and criticism as it took place on Commonwealth Day service at Westminster Abbey, which was attended by other senior members of the royal family. The following month, when asked a music-related question during a BBC1 interview for the Heads Together campaign, he commented, referring to the reports: "I've got into enough trouble with my dancing recently. So it's probably best to stay away from that one".
In July 2021, ITV aired a documentary entitled Harry and William: What Went Wrong?, centering on the princes' relationship. Before the film aired, footage of journalist Omid Scobie claiming that a senior aide from the Duke of Cambridge's household had briefed journalists about "concerns regarding Prince Harry's mental health" was partially cut from the documentary after Kensington Palace informed the broadcaster that the claim was "potentially defamatory".
Titles, styles, honours, and arms
Titles and styles
1982–2011: His Royal Highness Prince William of Wales
2011–present: His Royal Highness The Duke of Cambridge
in Scotland: 2011–present: His Royal Highness The Earl of Strathearn
The hereditary titles of Duke of Cambridge, Earl of Strathearn, and Baron Carrickfergus were announced on 29 April 2011 and formally patented on 26 May that year. Royal communications have stated that William is formally known as the Earl of Strathearn in Scotland. While some parts of the media have reported that William is known as Baron Carrickfergus in Northern Ireland, royal communications regarding William in Northern Ireland have referred to him as the Duke of Cambridge. He is a Royal Knight Companion of the Most Noble Order of the Garter (KG), a Knight of the Most Ancient and Most Noble Order of the Thistle (KT), a member of the Privy Council of the United Kingdom (PC), and a Personal Aide-de-Camp (ADC) to the Queen.
As a British prince, William does not use a surname for everyday purposes. For formal and ceremonial purposes, children of the Prince of Wales use the title "prince" or "princess" before their forename and follow it with their father's territorial designation. Thus, before becoming a duke when he married, Prince William was styled "Prince William of Wales". Such territorial designations are discarded by women when they marry and by men if they become peers in their own right, such as when Prince William was made a duke.
Although the name of the Royal House is Windsor, the surname Mountbatten-Windsor belongs to all the children and male-line descendants of Queen Elizabeth II and Prince Philip, and is used, if needed, by those who do not have the style of Royal Highness and the title Prince or Princess; when a female descendant marries, she traditionally takes her husband's surname from that point onward, and their children take their father's. Both Princes William and Harry used Wales as their surname for military purposes; this continues to be the case for William since his creation as Duke of Cambridge.
Military ranks
8 January 2006: Officer cadet
16 December 2006: Cornet (Second Lieutenant), The Blues and Royals (short service commission)
16 December 2006: Lieutenant, The Blues and Royals
1 January 2009: Captain, The Blues and Royals (and transferred to a full regular commission)
1 January 2016: Major
1 January 2008: Sub-lieutenant
1 January 2009: Lieutenant
1 January 2016: Lieutenant Commander
1 January 2008: Flying Officer
1 January 2009: Flight Lieutenant
1 January 2016: Squadron Leader
Honours
Prince William is the 1,000th member of the register of the Order of the Garter, and was officially invested by the Queen on 16 June 2008 at a service at St George's Chapel, Windsor Castle. The last time a monarch appointed a grandchild into the Order of the Garter was in 1894, when Queen Victoria invested Alfred, Hereditary Prince of Saxe-Coburg and Gotha.
23 April 2008: Royal Knight Companion of the Most Noble Order of the Garter (KG)
6 February 2002: Queen Elizabeth II Golden Jubilee Medal
25 May 2012: Extra Knight of the Most Ancient and Most Noble Order of the Thistle (KT)
6 February 2012: Queen Elizabeth II Diamond Jubilee Medal
30 October 2017: Tuvalu Order of Merit
6 February 2022: Queen Elizabeth II Platinum Jubilee Medal
Appointments
Since 6 July 2009: Honorary Bencher of the Honourable Society of the Middle Temple.
Since 23 June 2010: Royal Fellow of the Royal Society (FRS)
17 March 2013: Personal Aide-de-Camp to Her Majesty The Queen (ADC)
9 June 2016: Member of Her Majesty's Most Honourable Privy Council (PC)
Since 1 March 2017: Royal Honorary Fellow of the Royal Society of Edinburgh (HonFRSE)
Since 17 January 2018: Honorary Fellow of the Royal Society of Medicine (FRSM)
25 January 2020, 22 March 2021: Lord High Commissioner to the General Assembly of the Church of Scotland
Honorary military appointments
Canada
10 November 2009: Honorary Canadian Ranger
United Kingdom
Since 8 August 2006: Commodore-in-Chief of the Royal Navy Submarine Service
Since 8 August 2006: Commodore-in-Chief of Scotland
Since 3 October 2008: Honorary Air Commandant of RAF Coningsby
Since 10 February 2011: Colonel of the Irish Guards
Eponyms
Their Royal Highnesses The Duke and Duchess of Cambridge Award, University of Waterloo, Waterloo, Ontario
Duke and Duchess of Cambridge's Parks Canada Youth Ambassadors Program
Duke of Cambridge Public School, Bowmanville, Ontario
Arms
Ancestry
William is a member of the House of Windsor. Patrilineally, he descends from the House of Oldenburg, one of Europe's oldest royal houses; and more specifically the cadet branch known as the House of Glücksburg.
Through his mother, William descends from the Earls Spencer—a cadet branch of the Spencer family descended from the Earls of Sunderland; the senior branch are now also Dukes of Marlborough; the Barons Fermoy; and more anciently from Henry FitzRoy, 1st Duke of Grafton, and Charles Lennox, 1st Duke of Richmond—two illegitimate sons of King Charles II. As king, William would be the first monarch since Anne to undisputedly descend from Charles I and the first to descend from Charles II.
William descends matrilineally from Eliza Kewark, a housekeeper for his eighteenth-century ancestor Theodore Forbes—a Scottish merchant who worked for the East India Company in Surat. She is variously described in contemporary documents as "a dark-skinned native woman", "an Armenian woman from Bombay", and "Mrs. Forbesian". Genealogist William Addams Reitwiesner assumed Kewark was Armenian. In June 2013, BritainsDNA announced that genealogical DNA tests on two of William's distant matrilineal cousins confirm Kewark was matrilineally of Indian descent.
Bibliography
Books
HRH Prince William, "Introduction", in:
HRH The Duke of Cambridge, "Foreword", in:
Authored letters and articles
See also
Royal William, a German red rose named after Prince William shortly after his birth
Notes
References
Further reading
External links
The Duke of Cambridge at the official website of the British royal family
|-
|-
|-
|-
|-
|-
|-
1982 births
Living people
20th-century British people
21st-century British Army personnel
21st-century Royal Air Force personnel
21st-century Royal Navy personnel
Alumni of the University of St Andrews
Blues and Royals officers
British Anglicans
British environmentalists
British humanitarians
British male water polo players
British people of Indian descent
British people of American descent
British polo players
W
W
Graduates of the Royal Military Academy Sandhurst
Helicopter pilots
House of Windsor
Knights of the Garter
Knights of the Thistle
Members of the Privy Council of the United Kingdom
Mountbatten-Windsor family
People educated at Eton College
People educated at Ludgrove School
People educated at Wetherby School
Presidents of the Football Association
W
Cambridge, William
Royal Air Force squadron leaders
Royal Navy officers
Time 100 |
null | null | Olympic Oath | eng_Latn | The Olympic Oath (distinct from the Olympic creed) is a solemn promise made by one athlete, judge or official, and one coach at the Opening Ceremony of each Olympic Games. Each oath taker is from the host nation and takes the oath on behalf of all athletes, officials, or coaches at the Games. The athletes' oath was first introduced for the 1920 Summer Olympic Games, with oaths for the officials and coaches added in 1972 and 2010. The oath is usually said in the language of the nation which is hosting the games; however, in 1994 both the athletes' and officials' oaths were said in English. Until the 1984 games the oath takers swore upon their nation's flag; since then all have taken the oath whilst holding the Olympic Flag. All three of the oaths were combined into one beginning at the 2018 Winter Games.
The inspiration for an oath came from the Ancient Olympic Games where competitors swore on a statue of Zeus. An oath for the athletes was first thought of in 1906, following unsportsmanlike incidents. An athletes' oath was introduced for the 1920 games and Victor Boin was the first person to take the oath on behalf of all athletes. Giuliana Minuzzo was the first woman to take the athletes' oath at the winter games in 1956, while Heidi Schuller did likewise at the Summer Olympics in 1972. The first Olympic Champion to take the oath was Rudolf Ismayr, who took it at the 1936 Games. The only occasion where more than one person has said an oath occurred at the 1988 Summer Olympics when Hur Jae and Shon Mi-Na took the athletes' oath together, until 2021 when rules around gender equality decreed that each oath would be taken by a man and a woman. The oath has changed over the years to remove nationalism and to reflect drugs in sport and equality.
An oath for the officials was first discussed in the 1950s. It was not, however, until 1970 that the International Olympic Committee (IOC) voted to include an oath of the officials as well as athletes at the Olympic Games. The first oath for the officials was taken by Fumio Asaki at the 1972 Winter Olympic Games. When the Youth Olympics were created the IOC decided to have an oath for coaches as they realised that young athletes look to them particularly. This was introduced into the adult games for the 2012 edition.
History
An oath was an idea taken from the Ancient Olympic Games where competitors swore an oath beside a statue of Zeus. A call for an oath was announced as early as 1906 by International Olympic Committee (IOC) president and founder Pierre de Coubertin in the Revue Olympique (Olympic Review in French). This was done in an effort to ensure fairness and impartiality.
The Olympic Oath was first taken at the 1920 Summer Olympics in Antwerp by the fencer Victor Boin.
Boin's oath in 1920 was:
We swear. We will take part in the Olympic Games in a spirit of chivalry, for the honour of our country and for the glory of sport.
At a winter sports week in Chamonix in 1924, which were retrospectively call the Olympic Games in 1926, all the competitors took an Olympic style oath and were led by Camille Mandrillon. Rudolf Ismayr was the first Olympic Champion to take the oath, doing so at the 1936 Games in Berlin. In 1956 Giuliana Chenal-Minuzzo became the first woman to recite the oath.
In 1961, "swear" was replaced by "promise" and "the honour of our countries" by "the honour of our teams" in an effort to eliminate nationalism at the Olympic Games. Therefore the oath was as follows:
In the name of all competitors, I promise that we shall take part in these Olympic Games, respecting and abiding by the rules that govern them, in the true spirit of sportsmanship, for the glory of sport and the honour of our teams.
An oath for the officials had been discussed since the 1950s when the International Amateur Boxing Association asked all its officials to undertake an oath. In 1970 the IOC amended rule 57 of the charter and decreed that a judge from the host nation would also take an oath. The first judges' oath was taken at the 1972 Winter Olympics in Sapporo by Fumio Asaki and Heinz Pollay performed the task at that year's summer games in Munich.
The Munich games saw Heidi Schüller become the first female athlete to take the oath at the Summer Games; women had been competing at the games since 1900. At the 1988 Games the athletes' oath for the first time was undertaken by more than one person, when Hur Jae and Shon Mi-Na took the oath in unison.
All persons up until the 1980 Games swore the oath on their country's flag. In an attempt to eliminate nationalism from the oath ceremony all have sworn on the Olympic Flag since the 1984 games. The oaths are usually spoken in the language of the host nation (or athlete's mother tongue – Bojan Krizaj represented Yugoslavia but spoke Slovenian), but in 1994 both of the oaths were conducted in English rather than Norwegian.
In 1999, the IOC created the World Anti-Doping Agency (WADA) in an effort to form a more organized battle against doping. Thus, the Athletes Oath was amended to include references to doping and drugs. In 2010 the IOC recognised that younger athletes, in particular, are influenced by their coaches. At the first Youth Olympic Games, coaches were also required to undertake an oath for this reason. With effect from the 2018 Winter Olympics, the three oaths were combined into one, led by an athlete, although representatives of judges and coaches are still involved in the taking of the oath. In July 2021 the oath was further refined to highlight inclusion and equality. The changes to the wording came from the IOC Athletes’ Commission. In addition the number of oath takers at each games were expanded from three to six, to ensure a member of each gender was represented in each oath as part of a commitment to promote women in sport at all levels and in all structures, as stated in the Olympic Charter.
Athletes' Oath
The chosen athlete, a representative of all the participating Olympic competitors, recited the following from 2000 until 2016:In the name of all competitors, I promise that we shall take part in these Olympic Games, respecting and abiding by the rules which govern them, committing ourselves to a sport without doping and without drugs, in the true spirit of sportsmanship, for the glory of sport and the honor of our teams.
The Officials' Oath
The judge/official, also from the host nation, likewise held a corner of the flag and said the following:
In the name of all the judges and officials, I promise that we shall officiate in these Olympic Games with complete impartiality, respecting and abiding by the rules which govern them, in the true spirit of sportsmanship.
The Coaches' Oath
At the 2010 Summer Youth Olympics, an additional oath was taken by a coach; this was added to the protocol for the 2012 Games:
In the name of all the coaches and other members of the athletes' entourage, I promise that we shall commit ourselves to ensuring that the spirit of sportsmanship and fair play is fully adhered to and upheld in accordance with the fundamental principles of Olympism.
Unified Oath
Beginning in PyeongChang at the 2018 Winter Olympics, there was only one oath.
A representative for the athletes, judges, and coaches each recites the following lines respectively:
In the name of the athletes.
In the name of all judges.
In the name of all the coaches and officials.
The athletes' representative then completes the oath:
We promise to take part in these Olympic Games, respecting and abiding by the rules and in the spirit of fair play. We all commit ourselves to sport without doping and cheating. We do this, for the glory of sport, for the honour of our teams and in respect for the Fundamental Principles of Olympism.
Equality changes
In July 2021 the oath was updated to include elements around inclusion and equality. The oath from now on would be said by a man and a woman of each discipline.
The representatives for the athletes, judges, and coaches each recites the following lines respectively:
In the name of the athletes.
In the name of all judges.
In the name of all the coaches and officials.
The athletes' representative then completes the oath:
We promise to take part in these Olympic Games, respecting and abiding by the rules and in the spirit of fair play, inclusion and equality. Together we stand in solidarity and commit ourselves to sport without doping, without cheating, without any form of discrimination. We do this for the honour of our teams, in respect for the Fundamental Principles of Olympism, and to make the world a better place through sport.
Speakers
The athletes, judges and coaches that have delivered the Olympic Oath are listed below.
See also
Olympic symbols
References
External links
Recording of the Olympic Oaths being taken in 2012
Oath
Oaths |
null | null | Dallas Stars | eng_Latn | The Dallas Stars are a professional ice hockey team based in Dallas. They compete in the National Hockey League (NHL) as a member of the Central Division in the Western Conference, and were founded during the 1967 NHL expansion as the Minnesota North Stars, based in Bloomington, Minnesota. Before the 1978–79 NHL season, the team merged with the Cleveland Barons after the league granted them permission due to each team's respective financial struggles. Ultimately, the franchise relocated to Dallas for the 1993–94 NHL season. The Stars played out of Reunion Arena from their relocation until 2001, when the team moved less than into the American Airlines Center.
The Stars have won eight division titles in Dallas, two Presidents' Trophies as the top regular-season team in the league, the Western Conference championship three times, and the Stanley Cup in 1998–99. Joe Nieuwendyk won the Conn Smythe Trophy as the most valuable player of the playoffs that year.
In 2000, Neal Broten was inducted into the United States Hockey Hall of Fame. In 2009, Brett Hull became the first Dallas Stars player inducted into the Hockey Hall of Fame, followed by Ed Belfour and Joe Nieuwendyk in 2011 and Mike Modano in 2014; Modano is the highest-scoring player in franchise history. In 2010, brothers Derian and Kevin Hatcher were inducted to the United States Hockey Hall of Fame. The franchise has appeared in the Stanley Cup finals five times (1981, 1991, 1999, 2000, and 2020) and has won one championship (1999).
Franchise history
Minnesota North Stars (1967–1993)
The Minnesota North Stars began play in 1967 as part of the league's six-team expansion. Home games were played at the newly constructed Metropolitan Sports Center ("Met Center") in Bloomington, Minnesota. Initially successful both on the ice and at the gate, the North Stars fell victim to financial problems after several poor seasons in the mid-1970s.
In 1978, the North Stars merged with the Cleveland Barons (formerly the California Golden Seals), owned by George III and Gordon Gund. With both teams on the verge of folding, the league permitted the two failing franchises to merge. The merged team continued as the Minnesota North Stars, while the Seals/Barons franchise records were retired. However, the Gunds were the merged team's principal owners, and the North Stars assumed the Barons' place in the Adams Division in order to balance out the divisions. The merger brought with it a number of talented players, and the North Stars were revived—they reached the Stanley Cup Finals in 1981, where they lost in five games to the New York Islanders. However, by the early 1990s, declining attendance and the inability to secure a new downtown revenue-generating arena led ownership to request permission to move the team to the San Francisco Bay Area in 1990. The league rejected the request and instead agreed to award an expansion franchise, the San Jose Sharks, to the Gund brothers. The North Stars were sold to a group of investors that were originally looking to place a team in San Jose, although one of the group's members, former Calgary Flames part-owner Norman Green, would eventually gain control of the team. In the following season, the North Stars made it to the Stanley Cup Finals, only to lose to the Pittsburgh Penguins.
After the 1990–91 season, the North Stars suffered through declining profits coupled with distraction and uncertainty caused by relocation attempts. The team's fortunes were further impeded by the terms of the settlement with the Gund brothers, in which they were permitted to take a number of North Stars players to San Jose. Green explored the possibility of moving the team to Anaheim; however, the league decided instead to place the expansion Mighty Ducks there in 1993–94. In their final two seasons in Minnesota, the team adopted a new logo which omitted the "North" from "North Stars", leading many fans to anticipate the team heading south.
Relocation and early years in Dallas (1993–1998)
In 1993, amid further attendance woes and bitter personal controversy, Green obtained permission from the league to move the team to Dallas, for the 1993–94 season, with the decision announced on March 10, 1993. Green was convinced by former Dallas Cowboys quarterback Roger Staubach that Dallas would be a suitable market for an NHL team. He decided to retain the "Stars" nickname, as it matched Texas' official nickname as "The Lone Star State." The Stars would move into Reunion Arena, built in 1980, the downtown arena already occupied by the National Basketball Association's Dallas Mavericks.
To quell the ensuing controversy, the NHL promised that the Twin Cities would receive an expansion franchise in the future; that promise was fulfilled in 2000 in the form of the Minnesota Wild.
With the league changing the names of the conferences and divisions that season, the newly relocated Stars were placed in the Central Division of the Western Conference, and played their first game in Dallas on October 5, 1993, a 6–4 win against the Detroit Red Wings. In that game, Neal Broten scored the first Stars goal in Dallas. Dallas was an experiment for the league. At that time, the Stars would be one of the three southernmost teams in the league, along with the newly created Tampa Bay Lightning and Florida Panthers, as the league's first real ventures into southern non-traditional hockey markets. Though the Stars were relatively still unknown in the area, word of the team spread rapidly, and the immediate success of the team on the ice, as well as Mike Modano's career-best season (50 goals, 93 points) helped spur the team's popularity in Dallas. The Stars set franchise bests in wins (42) and points (97) in their first season in Texas, qualifying for the 1994 playoffs. The Stars further shocked the hockey world by sweeping the St. Louis Blues in the first round, but lost to the eventual Western Conference Champion Vancouver Canucks in the second round. The Stars' success in their first season, along with Modano's spectacular on-ice performances, would be an integral part of the Stars' eventual franchise success in the immediate years to come.
The almost immediate success of the Stars was also helped by the long legacy of minor-league hockey in the area. Both incarnations of the Central Hockey League had two teams in the area, the Fort Worth Texans and the Dallas Black Hawks for years before the Stars' arrival. These two teams were bitter rivals, and some of the traditions and famous rivalry incidents, including what is thought to be the first known use anywhere at an athletic event of "Rock and Roll Part 2" by Fort Worth and the famous "10 Cent Beer Night" near-riot in 1978 helped create awareness of hockey. Amateur and youth hockey were also extremely popular because of the long presence of the minor league teams.
Green, who had run into financial problems stemming from his business ventures outside of hockey, was forced to sell the team to businessman Tom Hicks in December 1995.
Tom Hicks Era and Building for a Championship
The 1994–95 season was shortened by an owners' lockout. The Stars traded captain Mark Tinordi along with Rick Mrozik to the Washington Capitals before the season began for Kevin Hatcher. Longtime North Stars hold-over Neal Broten was named his replacement, although he was traded too after only 17 games to the New Jersey Devils. Broten was replaced by Kevin's younger brother Derian Hatcher as team captain, a role he would serve in for the next decade. The Stars played only 48 games that season posting a record of 17–23–8. Despite the shortened season and the losing record, the Stars again made the playoffs, losing in five games to the Red Wings in the first round.
The 1995–96 season would be the first full season under new owner Tom Hicks. In the off-season, the Stars traded for former Montreal Canadiens' captain and three-time Frank J. Selke Trophy winner Guy Carbonneau, who was then with the St. Louis Blues. With the Stars struggling to begin the season, general manager and head coach Bob Gainey traded for center Joe Nieuwendyk from the Calgary Flames in exchange for Corey Millen and Jarome Iginla, then a Stars prospect. The Stars recorded only 11 wins in the first half of the season, and Bob Gainey relinquished his coaching duties in January to be the full-time general manager of the team. The Stars soon hired Michigan K-Wings head coach Ken Hitchcock to replace him; it would be his first NHL head coaching position. The Stars then traded for Benoit Hogue from the Toronto Maple Leafs late in the season, but ultimately finished in sixth place in the Central Division, missing the playoffs for the first time since moving to Texas.
In the 1996 off-season, the Stars continued to revamp their roster, adding defensemen Darryl Sydor from the Los Angeles Kings followed by Sergei Zubov from the Pittsburgh Penguins for Kevin Hatcher. Zubov would serve as the Stars' top defenseman and powerplay quarterback until leaving the league in 2009. On the ice, Ken Hitchcock's first season proved to be a good one. The Stars bested their 1994 totals, posting 48 wins and reaching the 100-point mark for the first time in franchise history. The Stars won the Central Division, their first division title since 1983–84 (when they were still the Minnesota North Stars) and were seeded second in the playoffs. Despite the regular-season success, the youthful Stars were upset in the first round by the Edmonton Oilers in seven games. Defenseman Grant Ledyard tripped in overtime of Game 7, allowing Todd Marchant to score the game- and series-winning goal on a breakaway against goaltender Andy Moog.
In the 1997 off-season, the Stars signed star goaltender Ed Belfour as a free agent after a well-publicized falling-out with the San Jose Sharks, which had traded a number of players to the Chicago Blackhawks to obtain him in January in the previous season. Andy Moog was allowed to leave via free agency, but later returned to the Stars as an assistant coach. The 1997–98 season was another banner year for the Stars. The Stars again set franchise records in wins (49) and points (109). Dallas acquired Mike Keane at the deadline from the New York Rangers. The Stars won the franchise's first Presidents' Trophy as the league's best regular-season team, as well as the Central Division title for the second season in a row. Belfour set franchise season records for goals against average (1.88), wins (37) and just missed out on the Jennings Trophy by one goal to Martin Brodeur of the New Jersey Devils. The Stars were the first overall seed for the 1998 Stanley Cup playoffs and defeated the eighth-seeded San Jose Sharks in six games in the first round. Notorious enforcer Bryan Marchment injured Joe Nieuwendyk's right knee, forcing him to miss the rest of the playoffs with torn ligaments. In the second round, they again met the Edmonton Oilers, this time defeating them in five games. However, without Nieuwendyk, the Stars lacked the firepower to overcome the defending Stanley Cup champion Detroit Red Wings in the Western Conference Finals and lost in six games. The Red Wings went on to win their second consecutive Stanley Cup.
Dominance, a Stanley Cup championship, and deep playoff runs (1998–2004)
In the 1998 off-season, after falling just short in the Western Conference Finals, the Stars added what they believed was the final piece toward winning a championship: star goalscoring winger Brett Hull. Hull had already had a successful career with the St. Louis Blues, with three consecutive 70-goal seasons and a Hart Memorial Trophy, but a fallout with Blues management led Hull to leave St. Louis via free agency. Additionally, this was the first season for the Stars in the Pacific Division after the 1998 NHL division re-alignment.
In the 1998–99 season, the Stars won 51 games, surpassing the 50-win mark for the first time in franchise history. They also recorded 114 points, which still stands today as a franchise record. They won the Pacific Division by 24 points, their third consecutive division title; a second consecutive Presidents' Trophy; the Jennings Trophy as the league's top defensive team; and were awarded home-ice advantage throughout the 1999 playoffs. Winger Jere Lehtinen was also awarded the Frank J. Selke Trophy.
In the first round of the playoffs, Dallas faced the Edmonton Oilers. The Stars swept the Oilers in four close games, winning Game 4 in the third overtime on a goal by Joe Nieuwendyk. They then faced the St. Louis Blues in the second round. After taking a 2–0 series lead, the Blues came back to tie the series. The Stars then won the next two games to beat the Blues in six games. The series again ended on an overtime goal, this time from Mike Modano. In the Conference Finals, they faced the Colorado Avalanche for the first time in Stars history. This would be the first of four playoff meetings between the Stars and Avalanche in the next seven years. After both the Stars and the Avalanche split the first four games at a 2–2 series tie, the Avalanche won Game 5 by a score of 7–5, taking a 3–2 series lead, The Stars rallied winning game six on the road, and Game 7 at home, both by 4–1 scorelines.
This was the Stars' first Stanley Cup Finals appearance as the Dallas Stars, although they made the finals twice as the Minnesota North Stars. They faced the Eastern Conference champion Buffalo Sabres, who had defeated the Toronto Maple Leafs 4–1 in the Eastern Conference Finals. After splitting the first four games, the Stars vaunted defense would hold the Sabres to only one goal in the next two, winning Game 5 2–0 and Game 6 2–1 on an overtime goal by Brett Hull. Hull's goal at 14:51 of the third overtime was allowed to stand only after a lengthy official review. That season, the league still had the "crease rule" in effect, which provided that if any player of the attacking team who did not have possession of the puck entered the crease before the puck, then any resulting goal was disallowed. Hull had initially gained possession of the puck outside the crease and had made a shot that was blocked by Buffalo goaltender Dominik Hasek. One of Hull's skates entered the crease as he corralled the rebound, and Hull's second shot scored the Cup-winning goal. The goal was eventually allowed, as having simply blocked Hull's shot rather than catching it, Hasek never took possession of the puck away from Hull. Officials therefore determined that rather than calling it a loose puck, Hull would be considered to have had continuous possession of the puck from before his first shot outside the crease. The complexity of the crease rule, and the attendant difficulties in understanding its application by fans and players alike, combined with the controversy arising out of the disputed Stanley Cup-winning goal, resulted in the crease rule being repealed the following season. Hull's goal marked the 13th time a Stanley Cup-winning goal was scored in overtime, and only the fourth to be scored in multiple overtimes. This was the only time between 1995 and 2003 that a team other than the New Jersey Devils, Colorado Avalanche or Detroit Red Wings won the Stanley Cup.
The team added veterans Kirk Muller, Dave Manson and Sylvain Cote in an effort to defend their Stanley Cup championship in . On December 31, 1999, Brett Hull scored his 600 and 601st career goals in a 5–4 win over the Mighty Ducks of Anaheim. The Stars ultimately won the Pacific Division for the second year in a row, and were seeded second in the Western Conference. Dallas then defeated the Edmonton Oilers and San Jose Sharks in the first and second rounds, both 4–1 series victories. The Stars, for the second season in a row, defeated the Colorado Avalanche in the Western Conference Finals in seven games to reach their second consecutive Stanley Cup Finals, where they met the New Jersey Devils. Because the Devils finished the regular season with one more point than Dallas, the Stars had to play their first playoff series without home-ice advantage since 1995. The Stars lost all three games at the Reunion Arena in the Finals, and lost the series in Game 6 on a double-overtime goal by New Jersey forward Jason Arnott.
Unsuccessful playoff runs
Hoping to win back the Stanley Cup, the Stars again captured the Pacific Division, posting a solid 48–24–8–2 record in the 2000–01 season. In the playoffs, the Stars met the Edmonton Oilers in the first round, battling back-and-forth through the first four games, with each game decided by one goal, including three going into overtime. Game 5 would also go to overtime, as the Stars took a 3–2 series lead on a goal by Kirk Muller. However, in Game 6 in Edmonton, the Stars did not need overtime, advancing to the second round with a 3–1 win. Facing the St. Louis Blues, the Stars would run out of gas, being swept in four straight games. The Game 2 loss would be the last NHL game played in Reunion Arena.
Moving into the brand new American Airlines Center for the 2001–02 season, the Stars got off to a slow start to the season, as goaltender Ed Belfour struggled through one of his worst seasons. Head coach Ken Hitchcock was eventually fired, being replaced by Rick Wilson. Despite the coaching change, the Stars continued to play poor hockey. With the prospect of missing the playoffs, the Stars traded 1999 Conn Smythe winner Joe Nieuwendyk and Jamie Langenbrunner to the New Jersey Devils in exchange for Randy McKay and Jason Arnott, who scored the game-winning goal in the 2000 Finals against Dallas. The Stars would eventually go on to post a respectable record of 35–28–13–5. However, it would not be enough for the playoffs, as they fell four points short of the final eighth spot in the Western Conference. Following the season, coach Rick Wilson would return to assistant coaching duties, as the Stars brought in Dave Tippett as his replacement.
As in the 2002 off-season, Ed Belfour left via free agency to the Toronto Maple Leafs. To begin the 2002–03 season, the Stars awarded the starting goaltending position to Marty Turco, who went on to have one of the best seasons in NHL history, posting the lowest goals-against average (GAA) since 1940, at 1.76. However, missing 18 games late in the season likely cost him a shot at the Vezina Trophy, awarded to the league's top regular-season goaltender. Regardless, the Stars posted the best record in the Western Conference at 46–17–15–4, and along the way, two-way star Jere Lehtinen won his third Frank J. Selke Trophy. In the playoffs, the Stars once again met the Edmonton Oilers, and once again the Oilers would prove a contentious opponent, winning two of the first three games. However, the Stars would prove the better team again by winning the next three games to take the series in six games. The Stars' second-round series against the Mighty Ducks of Anaheim got off to an unbelievable start, as the game went deep into overtime tied 3–3. However, Mighty Ducks goaltender Jean-Sebastien Giguere stopped 60 shots as the Ducks scored early in the fifth overtime. Game 2 would be more of the same, as the Ducks stunned the Stars in overtime. Desperately needing a win, the Stars bounced back to take Game 3 in Anaheim. However, the Ducks would take a 3–1 series lead by breaking a scoreless tie late in the third period of Game 4. In Game 5, the Stars finally solved Giguere by scoring four goals to keep their playoff hopes alive. However, the Stars' dreams of a return trip to the Stanley Cup Finals would end in heartbreaking fashion as the Ducks broke a 3–3 tie with 1:06 left in Game 6 on a goal by defenseman Sandis Ozolinsh.
Coming off their disappointing playoff loss, the Stars would get off to a shaky start to the 2003–04 season, as they played mediocre hockey through the first three months of the season, posting a sub-.500 record. However, as the calendar turned to 2004, the Stars began to find their game, as they posted a 9–4–3 record in January. As the season wore on, the Stars would get stronger, climbing up the playoff ladder and eventually reaching second place in the Pacific Division, where they finished with a solid 41–26–13–2 record; Marty Turco had another outstanding season, recording a 1.98 GAA. However, the Stars could not carry their momentum into the playoffs, as they were beaten by the Colorado Avalanche in five games in the first round.
Post-lockout era (2005–2008)
Coming out of the owners' lockout that cancelled the entire 2004–05 season, the Stars remained one of the strongest teams in the Western Conference for the start of 2005–06, as they won four of their first five games on the way to a solid October. November would be even better for Dallas, as they won 10 of 13 games and took over first place in the Pacific Division, a position they would hold most of the season, as they went on to finish with a terrific record of 53–23–6. One reason for the Stars' success was their strong play in shootouts, as forward Jussi Jokinen was nearly automatic, making 10-of-13 shot attempts. Also performing strongly in shootouts was Sergei Zubov, who used a slow-but-steady backhand to go 7-for-12, as the Stars ultimately won 12 of 13 games that were settled by a shootout. As the number two seed in the Western Conference, the Stars faced the seventh-seeded Colorado Avalanche. The Stars were favorited to win the Western Conference, and some even predicted them to win the Stanley Cup. However, the Stars would stumble right from the start, losing Game 1 by a score of 5–2 as the Avalanche scored five unanswered goals after the Stars jumped out to a promising 2–0 lead. Game 2 would see the Stars suffer another setback at home, as the Stars lost in overtime 5–4 on a goal by Joe Sakic. On the road in Game 3, the Stars led 3–2 in the final minute before the Avalanche forced overtime on a goal by Andrew Brunette, while Alex Tanguay won the game just 69 seconds into overtime to put the Stars in a 3–0 hole. The Stars would avoid the sweep with a 4–1 win in Game 4, but overtime would doom them again in Game 5, as Andrew Brunette scored the series winner at 6:05, ending the Stars' playoffs hopes after just five games.
Dallas hosts the All-Star Game (2006–07 season)
Following the previous season's disappointing first-round playoff upset at the hands of the seventh-seeded Avalanche, the Stars made a number of changes during the 2006 off-season. Former Stars goalkeeper Andy Moog was promoted to assistant general manager for player development (he kept his job as goaltending coach) and former player Ulf Dahlen was hired as an assistant coach. The Stars allowed center Jason Arnott, defenseman Willie Mitchell and goaltender Johan Hedberg to leave as free agents. Forward Niko Kapanen was traded to the Atlanta Thrashers and the remaining two years on fan-favorite right winger Bill Guerin's contract were bought out. The Stars also received Patrik Stefan and Jaroslav Modry in the Atlanta trade, and signed Eric Lindros, Jeff Halpern, Matthew Barnaby and Darryl Sydor as free agents. Young goaltender Mike Smith was promoted to the NHL to serve as Marty Turco's backup. During the season, key future pieces – center Mike Ribeiro and defenseman Mattias Norstrom – were added through separate trades. Young players Joel Lundqvist, Krys Barch, Nicklas Grossmann and Chris Conner all saw significant ice time while other players were out of the lineup with injuries.
On September 29, 2006, Brenden Morrow was announced as new team captain, taking the role over from Mike Modano, who had served as the incumbent since 2003. On March 13, 2007, Modano scored his 500th career NHL goal, making him only the 39th player and second American to ever reach the milestone. On March 17, Modano scored his 502nd and 503rd NHL goals, breaking the record for an American-born player, previously held by Joe Mullen.
On January 24, 2007, the NHL All-Star Game was held at the American Airlines Center. Defenseman Philippe Boucher and goaltender Marty Turco would represent the Stars as part of the Western Conference All-Star roster. The Stars qualified for the playoffs as the sixth seed in the Western Conference and squared off against the Vancouver Canucks in the first round. Marty Turco delivered three shutout wins in Games 2, 5 and 6, but the Stars' offense failed to capitalize and they lost the series in seven games, the third season in a row that they lost in the first round.
Return to playoff prominence (2007–08 season)
In the 2007 NHL Entry Draft, the Stars drafted the relatively unknown Jamie Benn 129th overall. After starting a lackluster 7–7–3 in the 2007–08 season, general manager Doug Armstrong was fired by the team. He was replaced by an unusual "co-general manager" arrangement of former assistant GM Les Jackson and former Stars player Brett Hull. On November 8, 2007, Mike Modano became the top American born point scorer of all-time, finishing off a shorthanded breakaway opportunity on San Jose Sharks goaltender Evgeni Nabokov.
On February 26, 2008, just hours before the trade deadline, the Stars traded for All-Star center Brad Richards from the Tampa Bay Lightning in exchange for backup goaltender Mike Smith and forwards Jussi Jokinen and Jeff Halpern. The Stars rallied to a final record of 45–30–7 and qualified for the playoffs as the fifth seed, matching up with the defending Stanley Cup champion, the Anaheim Ducks, in the first round. After a rough end to the season, only winning two games in March of that year, the Stars shocked everyone by winning the first two games of the series in Anaheim, and then would go on to finish off the Ducks in six games, their first playoff series win since 2003. In the second round, the Stars matched up with the Pacific Division champion San Jose Sharks. Once again, the Stars surprised everyone by winning the first two games of the series on the road. In Game 2, Brad Richards tied an NHL record by recording four points in the third period. The Stars would then take a 3–0 lead after a Mattias Norstrom overtime goal in Game 3. Captain Brenden Morrow finished the Sharks off in Game 6 with a powerplay goal nearly halfway into the fourth overtime. The win sent the Stars to their first Conference Finals since 2000, where they met the powerhouse Detroit Red Wings. After falling behind 3–0, the Stars made a series of it, however, winning two games before ultimately finally being ousted by the Red Wings in six games.
Transition seasons (2008–2011)
The 2008–09 season saw the early loss for the season of captain Brenden Morrow to an ACL tear. Off-season free agent acquisition Sean Avery caused a media uproar over comments he made to a Canadian reporter about ex-girlfriend Elisha Cuthbert and her relationship with Calgary Flames' defenseman Dion Phaneuf before a game in Calgary. The incident forced the team to suspend Avery for the season; he was later waived by the Stars. That incident, plus injuries to key players Brad Richards and Sergei Zubov, caused the Stars to tailspin to a 12th-place finish and the first missed playoffs for Dallas since 2002.
In the wake of the season, the Stars hired a new general manager, former player and alternate captain Joe Nieuwendyk. Hull and Jackson remained with the Stars, but were reassigned to new roles within the organization. Less than a week after he was hired, Nieuwendyk fired six-season head coach Dave Tippett on June 10, 2009, and hired Marc Crawford the next day as his replacement. Other off-season moves included the addition of Charlie Huddy as assistant coach in charge of defense and the promotions of Stu Barnes and Andy Moog to assistant coaches.
The Stars' 2009–10 season was similar to the previous one. Inconsistent play and defensive struggles plagued the team throughout the season, as they failed to adjust to Marc Crawford's new offensively-minded system, and owner Tom Hicks' financial troubles prevented the team from spending more than $45 million on payroll, over $11 million beneath the league salary cap. The Stars failed to win more than three games in a row all season, finished in last in the Pacific Division and repeated their 12th place conference finish from the year before with a record of 37–31–14 for 88 points. This was the first time that they would miss the playoffs two seasons in a row since the Stars moved to Texas. In the off-season, long-time goaltender Marty Turco was let go in favor of Kari Lehtonen to become the team's number one goaltender for the future. In the Star's last game of the season, away against the Minnesota Wild, Mike Modano was named the game's first star and skated around the rink after the game wearing his North Stars uniform, receiving a rousing ovation.
In the 2010 off-season, the Stars released Marty Turco and Mike Modano, the face of the franchise for the past two decades. Modano subsequently signed with the Detroit Red Wings while Turco joined the Chicago Blackhawks. Winger Jere Lehtinen, who played his entire career with the Stars, announced his retirement in December 2010. The team also made key acquisitions, such as winger Adam Burish (who was on the 2010 Stanley Cup champion Chicago Blackhawks) and goaltender Andrew Raycroft. They also gave Jonathan Cheechoo a try-out, but he was cut and later signed with division rivals San Jose Sharks.
To begin the 2010–11 season, the Stars won their first three games in a row, going on a three-game win streak for the first time since the 2008–09 season by beating the New Jersey Devils in overtime in the season opener, New York Islanders in a shootout and against Mike Modano and the Red Wings. In that same game in the home opener against the Red Wings, the Stars crowd gave Modano a standing ovation as he was shown on the American Airlines Center jumbotron during a timeout in the game.
After a hot start to the new season, the Stars would dominate the first half of the season, staying in both the Pacific Division lead and within the top three spots of the Western Conference. It seemed like the old Stars were back, as through the first half of the season they compiled a 30–15–6 record. However, after the All-Star Break the Stars went into a slump, going on numerous losing streaks which included one-goal losses and blowing late leads in numerous games. Through this though, the Stars still remained in the playoff picture. On the day of the trade deadline, the Stars traded up and comer James Neal and Matt Niskanen to the Pittsburgh Penguins in exchange for defenseman Alex Goligoski. After struggling much of the second half of the season the Stars still had a chance to make the playoffs by winning all their games in the month of April. They won all of them except for their season finale as they lost on the road at the Minnesota Wild 5–3, costing them a playoff spot.
After missing the playoffs for the third consecutive season, Dallas fired coach Marc Crawford on April 12, 2011. In the spring of 2011, according to Darren Dreger of TSN, the team had been "financially managed" by the league for over a year. On June 16, 2011, Dallas hired Glen Gulutzan to be head coach, making him the sixth coach since the franchise's move from Minnesota. On September 13, 2011, lenders voted to agree to have the Stars file for bankruptcy and sold at auction. On September 21, 2011, Mike Modano announced his retirement from the league. By October 22, 2011, competing bids to buy the club were due. Vancouver businessman and Kamloops Blazers owner Tom Gaglardi's bid was the only one submitted, clearing the way for him to enter the final stages of taking over ownership of the team. Gaglardi's purchase was approved by the NHL Board of Governors on November 18, 2011.
A bankruptcy court judge approved the bid for an enterprise value of $240 million. First lien creditors got about 75 cents on the dollar. The Stars lost $38 million during their last fiscal year and $92 million over the last three seasons.
Tom Gaglardi era and Finals trip (2011–present)
As the new owner, Gaglardi's first move was bringing back former Stars President Jim Lites to once again take the reins as team president and CEO. To begin the 2011–12 season, the Stars once again jumped out to a fast start, going 22–15–1 through the first 38 games of the season. However, when the second half of the season began, the Stars slumped through the months of January and February, before getting hot again in late February. Throughout March, the Stars regained the lead of the Pacific Division. Beginning on March 26, 2012, the Stars embarked on a western road trip that saw them visit the Calgary Flames, Edmonton Oilers, Vancouver Canucks and San Jose Sharks. Going into the road trip, the Stars were in control of their own destiny, having to gain four points on the road trip to win their first Pacific Division title since the 2005–06 season. After the Stars lost 5–4 in Calgary to the Flames, the Stars beat the Oilers two nights later, 3–1. This would be their last win of the season, as the Stars were rolled over by the Canucks and Sharks. Even though they had lost the division crown, the Stars still had a chance to clinch a playoff spot. However, the Stars were eliminated from playoff contention on April 5 in a 2–0 loss to the playoff-bound Nashville Predators. The team failed to qualify for the playoffs for the fourth consecutive year, setting a franchise record for futility.
On July 1, 2012, the team signed free agent veterans Ray Whitney, Aaron Rome and Jaromir Jagr. The next day, the Stars traded fan-favorite Steve Ott and Adam Pardy to the Buffalo Sabres in exchange for center Derek Roy. When 2012–13 NHL lockout ended, the Stars began an up-and-down season, although staying in the race for one of the eight playoff spots in the shortened season, which had only 48 games. In mid-season, forward Michael Ryder was traded to his former team, the Montreal Canadiens, for Erik Cole. This shocked many Stars fans, as Ryder was a fan-favorite to Stars fans and had been producing good stats for the Stars throughout the shortened season. Before the trade deadline in early April, the Stars began to falter, and the team's captain, Brenden Morrow, was traded to the Pittsburgh Penguins; Roy was traded to the Vancouver Canucks; Jagr to the Boston Bruins; and Tomas Vincour to the Colorado Avalanche to close out the NHL trade deadline, all in exchange for draft picks and prospects. After all of the trades at the trade deadline, many counted the Stars out, as it seemed they were beginning to rebuild and throw in the towel during the season. However, the Stars' remaining young players pulled together to win six of their next eight games, thus propelling the Stars back into the 2013 playoff race. The Stars soon became the new Cinderella team and were getting better as their newfound success went on. However, the Stars dropped their final five games, losing all of them and gaining only one point in their final five games, which eliminated them from playoff contention. The Stars had now missed the playoffs for five straight seasons, continuing to set the all-time record in the franchise (dating back to the Minnesota days) for most consecutive seasons without a playoff appearance.
The day after their final regular-season game of the 2012–13 season (a 3–0 loss to the Detroit Red Wings), the Stars fired general manager Joe Nieuwendyk. The next day, the Stars introduced their 11th all-time general manager, Jim Nill, the former assistant general manager of the Red Wings. On May 14, 2013, the coaching staff was also fired, and on May 31, 2013, Scott White was re-introduced as the director of hockey operations. Nill made his first big trade as general manager when he acquired Tyler Seguin, Rich Peverley and Ryan Button from the Boston Bruins in exchange for Loui Eriksson, Reilly Smith, Matt Fraser and Joe Morrow. Under new head coach Lindy Ruff, the Stars made it to the 2014 playoffs with a successful run on a 40–31–11 record, finishing fourth in the new-look Central Division for the 2013–14 season. They lost in the first round of the playoffs to the Anaheim Ducks in six games with a 4–2 Game 6 loss.
Nill made another big trade as general manager when he acquired Jason Spezza and Ludwig Karlsson from the Ottawa Senators in exchange for Alex Chiasson, Nick Paul, Alex Guptill and a second-round pick in the 2015 NHL Entry Draft. He also signed Senators player Ales Hemsky as a free agent on a three-year, $12 million contract on July 1, 2014. Despite these moves, the Stars finished with the second-lowest goaltender save percentage in the league during the 2014–15 season, which resulted in them failing to qualify for the 2015 playoffs due to their sixth-place finish in the Central Division. The lone bright spot of the 2014–15 season was Jamie Benn winning the Art Ross Trophy. On April 11, 2015, Benn scored four points in the Stars' last regular-season game to finish with 87 points on the season and win the Art Ross Trophy. His final point, a secondary assist with 8.5 seconds left in the game, allowed him to overtake John Tavares for the award.
In the 2015–16 season, the Stars won their first Central Division title since 1998 and posted the best regular-season record in the Western Conference. In the first round of the playoffs, they defeated the Minnesota Wild in six games. In the second round, they faced the St. Louis Blues, for the first time since 2001, and lost the series in seven games. The Stars finished with a 34–37–11 record in an injury-plagued 2016–17 season, missing the playoffs for the seventh time in the past nine seasons. As a result, the team announced that head coach Lindy Ruff's contract would not be renewed. Ken Hitchcock returned as head coach for the following season, and once again, the Stars missed the playoffs. Hitchcock retired after the season and was succeeded at the head coaching spot by Jim Montgomery. Montgomery's first season as the Stars coach saw the team make the playoffs, defeating the Nashville Predators in a six-game series the first round, but falling to the eventual Stanley Cup champion St. Louis Blues in seven games, losing game seven on the road in double-overtime by a score of 2–1 despite a 52-save performance by goaltender Ben Bishop.
In the 2019–20 season, the Stars defeated the Vegas Golden Knights in five games to advance to the Stanley Cup Finals for the first time since 2000. They went on to lose to the Tampa Bay Lightning in six games. In the next season, they missed the playoffs for the first time since 2018.
Team information
Arena
When the Stars first moved from Minnesota, they played in Reunion Arena, which they shared with the National Basketball Association's Dallas Mavericks. For hockey, Reunion Arena held 17,001 for NHL games. Throughout the hockey history of Reunion Arena, the arena was known for having one of the worst ice surfaces in the league, especially in its final days hosting the Stars. The Stars played at Reunion for eight years, from 1993–2001. Before the 2001–02 NHL season, both the Stars and the Mavericks moved into the new American Airlines Center, which is in the Victory Park neighborhood of Dallas, just north of Reunion Arena. The American Airlines Center holds 18,584 for Stars and NHL games. On January 24, 2007, the AAC hosted the 2007 National Hockey League All-Star Game. The AAC and the Dallas Stars have won several local and NHL awards for the "Best Fan Experience".
Broadcast
All Dallas Stars games are broadcast on radio on KTCK under a five-year deal announced in January 2009. KTCK replaced WBAP, which had broadcast games since the beginning of the 1994 season after KLIF has broadcast the first season in Dallas in 1993. Television coverage occurs primarily on Bally Sports Southwest, with KTXA (Channel 21) or BSSW+ broadcasting games when BSSW has a conflict.
The Stars, along with the Buffalo Sabres and Carolina Hurricanes are one of only three NHL teams to simulcast the entirety of their games on TV and radio, which the team has done since their 1993 arrival in Dallas. The original broadcast team from 1993 to 1996 was Mike Fornes (play-by-play) and Ralph Strangis (color). Fornes left the broadcast team after the 1995–96 season; Strangis moved to the play-by-play role and color commentator Daryl "Razor" Reaugh was added. Although both the DFW area's large media market and the team's fan base could theoretically support separate television and radio broadcast teams, the Stars have continued simulcasting due to the popularity of "Ralph and Razor" (as they are known) among local listeners and viewers. Like other NHL teams, the Stars now have a live radio broadcast transmitted inside American Airlines Center on 97.5 FM. This is done because AM radio signals often cannot penetrate concrete and steel building exteriors.
Strangis retired from the booth after the 2014–15 season and was replaced by Dave Strader. In June 2016, Strader was diagnosed with cholangiocarcinoma, a fairly rare and aggressive form of cancer of the bile duct. To begin the 2016–17 season, Reaugh assumed play-by-play duties while Strader underwent treatment. Studio analyst and former Stars defenseman, Craig Ludwig, took over as color commentator. During a break in Strader's treatment, he returned to the broadcast booth on February 18, 2017, a 4–3 overtime home win against the Tampa Bay Lightning. After the game, the Stars saluted Strader at center ice. On October 1, 2017, Strader died of cancer at age 62, leaving Reaugh and Ludwig to call the 2017–18 season. In July 2018, the Stars announced that Reaugh would return to color commentary for the 2018–19 season, with former studio host Josh Bogorad taking over the play-by-play.
Logo and jersey design
When they debuted in Dallas for the 1993–94 season, they kept the same uniform design from their later Minnesota days, except for the addition of the Texas logo patch on the shoulders. Away uniforms were black and home uniforms were white. With minor trim changes, a darker shade of green, and the word 'Dallas' added in the 1994–95 season, they kept this design until their 1999 Stanley Cup-winning campaign. The black pants have the word "Dallas" in gold run through the sides with green stripes.
In the 1997–98 season, the Stars introduced an alternate uniform that partly resembled those worn during the NHL All-Star Game at the time. The uniform was mostly green on top and black at the bottom, in a star-shaped design. For the 1999–2000 season, it became the primary away uniform, and was paired with a new home uniform featuring the same basic design, with white on top and green at the bottom. They kept this design until the 2006–07 season, during which the league switched color designations on home and away jerseys in the 2003–04 season. The striping was also eliminated on the black pants.
The Stars introduced an alternate jersey for the 2003–04 season that proved both embarrassing and unpopular to critics and fans. The uniform, which was black with a green bottom and red trim, featured a modern representation of the constellation Taurus topped by a trailing green star with red trail marks. However, fans and critics derided the uniform crest for its resemblance to a uterus, nicknaming it the "Mooterus". The uniform was used until the 2005–06 season.
With the switch to the Reebok Edge uniform system, the jerseys underwent a complete redesign. The home black jersey, introduced for the 2007–08 season, features the player's number on the chest and an arched 'Dallas' in white with gold trim, with the primary logo on the shoulders. The primary away jerseys, which were used from 2007 to 2010 and was used as an alternate for its final season, had the Stars logo crest in front and the uniform number on the top right, with the Texas alternate logo on the shoulders. An alternate white jersey based on the home black jersey was introduced for the 2008–09 season; they became the regular away uniforms for the 2010–11 season. The lettering is in green with gold and black trim. Both uniforms were used until the 2012–13 season.
A new logo and uniforms were introduced for the 2013–14 season. Silver replaced gold as the tertiary color, while green (in a bright new shade called "Victory Green", similar to the old North Stars' shade of green) was reintroduced as a primary uniform color. The new logo features the letter D centering a star, symbolizing Dallas' nickname as "The Big D". The home uniforms are in green with black and white striping, while the away uniforms are in white with a green shoulder yoke, and black and green striping. The inner collar features the team name on the home uniforms, and the city name on the away uniforms. The secondary logo, with the primary inside a roundel with the team name, is featured in the shoulders.
The Stars unveiled a special edition uniform for the 2020 NHL Winter Classic, featuring a design inspired from the defunct USHL Dallas Texans hockey club. The uniform features the word "Stars" (with the star substituting for the letter A) in the middle of the big letter "D". Both are in white with silver trim. A silhouette of the state of Texas in silver and green trim with the letter "D" inside adorns the left sleeve.
For the 2020–21 season, the Stars added a black alternate uniform with neon green accents, featuring the alternate "Texas D" logo in front. A neon green star is added on the left side of the collar, symbolizing Texas' "Lone Star State" nickname. Numbers are in 3D-accented black with neon green accents. The Stars also released a "Reverse Retro" alternate uniform in collaboration with Adidas, bringing back the 1999–2007 uniform design but with the current color scheme in place of the original green, gold and black colors. The design also featured white pants.
Minor league affiliates
The Dallas Stars are presently affiliated with two minor league teams. The Texas Stars are the American Hockey League (AHL) affiliate of the Dallas Stars, who after becoming unaffiliated with the Iowa Stars in 2008, did not have an AHL affiliate for the 2008–09 season. The Texas Stars began play in the AHL in the 2009–10 season as an affiliate for Dallas. They are located in Cedar Park, Texas (northwest of Austin).
In addition to their AHL affiliate, the Dallas Stars are also affiliated with the Idaho Steelheads of the ECHL. Based in Boise, the Steelheads have played home games in CenturyLink Arena Boise since 2003.
Traditions
Since the 2005–06 season, the national anthems are performed by Celena Rae, a Fort Worth native and former semi-finalist on American Idol. During "The Star-Spangled Banner", the fans yell the word "stars" in the lines "whose broad stripes and bright stars" and "O say does that star spangled banner yet wave". At games, as part of the entertainment, a Kahlenberg KDT-123 fog horn sounds after every Stars goal.
When the Stars take the ice at the beginning of each game, the song "Puck Off" (also referred to as the "Dallas Stars Fight Song") by Pantera, is played in the arena. Members of the Dallas–Fort Worth–Arlington metropolitan area band had become friends with members of the Stars in the 1990s, especially following the team's Stanley Cup win in 1999. "Puck Off" later also became the team's goal celebration song. During the song, fans chant the only lyrics in the song, "Dallas! Stars! Dallas! Stars!" while pumping their fists in the air. After a penalty on the opposing team is announced, Stars public address announcer shouts, "Your Dallas Stars are on...the..." with the fans finishing "power play!", immediately after which a clip of the Muse song "Knights of Cydonia" is played. After each Stars win, the Stevie Ray Vaughan (a Dallas native) and Double Trouble song "The House is a Rockin'" is played. A song of unknown origin called "The Darkness Music" is played after nearly every away goal. In recent years, fans have usually yelled, "Who cares!?" after away team goals are announced by the public address announcer.
The Stars have hosted a home game on New Year's Eve every year since 1997, except in 2004, 2012, 2019, 2020 and 2021 (the first two being due to lockouts, the third being due to their participation in the 2020 NHL Winter Classic, the fourth being due to the delayed offseason caused by the COVID-19 pandemic, and the fifth was postponed to February 13 due to a team-wide COVID-19 outbreak). For a period of time, the game coincided with the Big D NYE celebration (now on hiatus) on the south side of the arena in AT&T Plaza.
Season-by-season record
This is a partial list of the last five seasons completed by the Stars. For the full season-by-season history, see List of Dallas Stars seasons.
Note: GP = Games played, W = Wins, L = Losses, OTL = Overtime losses, Pts = Points, GF = Goals for, GA = Goals against
Players
Current roster
Team captains
Note: This list does not include former captains of the Minnesota North Stars and Oakland Seals
Mark Tinordi, 1993–1995
Neal Broten, 1995
Derian Hatcher, 1995–2003
Mike Modano, 2003–2006
Brenden Morrow, 2006–2013
Jamie Benn, 2013–present
League and team honors
Awards and trophies
Stanley Cup
1998–99
Presidents' Trophy
1997–98, 1998–99
Clarence S. Campbell Bowl
1998–99, 1999–2000, 2019–20
Art Ross Trophy
Jamie Benn: 2014–15
Conn Smythe Trophy
Joe Nieuwendyk: 1998–99
Frank J. Selke Trophy
Jere Lehtinen: 1997–98, 1998–99, 2002–03
Lester Patrick Trophy
Neal Broten: 1997–98
Roger Crozier Saving Grace Award
Ed Belfour: 1999–2000
Marty Turco: 2000–01, 2002–03
William M. Jennings Trophy
Ed Belfour and Roman Turek: 1998–99
First-round draft picks
Note: This list does not include selections of the Minnesota North Stars.
1993: Todd Harvey (9th overall)
1994: Jason Botterill (20th overall)
1995: Jarome Iginla (11th overall)
1996: Ric Jackman (5th overall)
1997: Brenden Morrow (25th overall)
1998: None
1999: None
2000: Steve Ott (25th overall)
2001: Jason Bacashihua (26th overall)
2002: Martin Vagner (26th overall)
2003: None
2004: Mark Fistric (28th overall)
2005: Matt Niskanen (28th overall)
2006: Ivan Vishnevskiy (27th overall)
2007: None
2008: None
2009: Scott Glennie (8th overall)
2010: Jack Campbell (11th overall)
2011: Jamie Oleksiak (14th overall)
2012: Radek Faksa (13th overall)
2013: Valeri Nichushkin (10th overall)
2013: Jason Dickinson (29th overall)
2014: Julius Honka (14th overall)
2015: Denis Guryanov (12th overall)
2016: Riley Tufte (25th overall)
2017: Miro Heiskanen (3rd overall) and Jake Oettinger (26th overall)
2018: Ty Dellandrea (13th overall)
2019: Thomas Harley (18th overall)
2020: Mavrik Bourque (30th overall)
2021: Wyatt Johnston (23rd overall)
Hall of Famers
The Dallas Stars presently acknowledge an affiliation with a number of inductees to the Hockey Hall of Fame. Hall of Fame inductees acknowledged by the Stars include 13 former players and three builders of the sport. The three individuals recognized as builders by the Hall of Fame includes former Stars executives, and head coaches. The tenure of five player inductees, and the three builders acknowledged by the Stars occurred when the franchise was based in Minnesota (1967–1993). Mike Modano is the only Hall of Fame inductee that played with the franchise before and after its relocation to Dallas.
In addition to players and builders, the Stars also acknowledge an affiliation Hockey Hall of Fame inductee Dave Strader. Providing play-by-play for the Stars, he was a recipient of the Hall of Fame's Foster Hewitt Memorial Award for his contributions in hockey broadcasting.
Retired numbers
Notes:
The NHL retired Wayne Gretzky's No. 99 for all its member teams at the 2000 NHL All-Star Game.
Franchise scoring leaders
These are the top-ten point-scorers in franchise (Minnesota and Dallas) history. Figures are updated after each completed NHL regular season.
– current Stars player
''Note: Pos = Position; GP = Games Played; G = Goals; A = Assists; Pts = Points; P/G = Points per game
Franchise individual records
Most goals in a season: Dino Ciccarelli; Brian Bellows, 55 (1981–82; 1989–90)
Most assists in a season: Neal Broten, 76 (1985–86)
Most points in a season: Bobby Smith, 114 (1981–82)
Most penalty minutes in a season: Basil McRae, 378 (1987–88)
Most points in a season, defenseman: Craig Hartsburg, 77 (1981–82)
Most points in a season, rookie: Neal Broten, 98 (1981–82)
Most goals in a season, rookie: Neal Broten, 38 (1981–82)*
Most wins in a season: Marty Turco, 41 (2005–06)
Most shutouts in a season: Marty Turco, 9 (2003–04)
See also
List of Dallas Stars head coaches
List of Dallas Stars players
List of Minnesota North Stars/Dallas Stars general managers
Notes
References
External links
National Hockey League teams
1967 establishments in Minnesota
Central Division (NHL)
Ice hockey clubs established in 1967
Ice hockey teams in Texas
National Hockey League in the Sun Belt |
null | null | St. Peter's Basilica | eng_Latn | The Papal Basilica of Saint Peter in the Vatican (), or simply Saint Peter's Basilica (), is a church built in the Renaissance style located in Vatican City, the papal enclave that is within the city of Rome. It was initially planned by Pope Nicholas V and then Pope Julius II to replace the aging Old St. Peter's Basilica, which was built in the 4th century by Roman emperor Constantine the Great. Construction of the present basilica began on 18 April 1506 and was completed on 18 November 1626.
Designed principally by Donato Bramante, Michelangelo, Carlo Maderno and Gian Lorenzo Bernini, St. Peter's is the most renowned work of Renaissance architecture and the largest church in the world by interior measure. While it is neither the mother church of the Catholic Church nor the cathedral of the Diocese of Rome (these equivalent titles being held by the Archbasilica of Saint John Lateran in Rome), St. Peter's is regarded as one of the holiest Catholic shrines. It has been described as "holding a unique position in the Christian world" and as "the greatest of all churches of Christendom."
Catholic tradition holds that the basilica is the burial site of Saint Peter, chief among Jesus's apostles and also the first Bishop of Rome (Pope). Saint Peter's tomb is supposedly directly below the high altar of the basilica, also known as the Altar of the Confession. For this reason, many popes have been interred at St. Peter's since the Early Christian period.
St. Peter's is famous as a place of pilgrimage and for its liturgical functions. The pope presides at a number of liturgies throughout the year both within the basilica or the adjoining St. Peter's Square; these liturgies draw audiences numbering from 15,000 to over 80,000 people. St. Peter's has many historical associations, with the Early Christian Church, the Papacy, the Protestant Reformation and Catholic Counter-reformation and numerous artists, especially Michelangelo. As a work of architecture, it is regarded as the greatest building of its age. St. Peter's is one of the four churches in the world that hold the rank of Major papal basilica, all four of which are in Rome, and is also one of the Seven Pilgrim Churches of Rome. Contrary to popular misconception, it is not a cathedral because it is not the seat of a bishop; the cathedra of the pope as Bishop of Rome is at Saint John Lateran.
Overview
St. Peter's is a church built in the Renaissance style located in the Vatican City west of the River Tiber and near the Janiculum Hill and Hadrian's Mausoleum. Its central dome dominates the skyline of Rome. The basilica is approached via St. Peter's Square, a forecourt in two sections, both surrounded by tall colonnades. The first space is oval and the second trapezoidal. The façade of the basilica, with a giant order of columns, stretches across the end of the square and is approached by steps on which stand two statues of the 1st-century apostles to Rome, Saints Peter and Paul.
The basilica is cruciform in shape, with an elongated nave in the Latin cross form but the early designs were for a centrally planned structure and this is still in evidence in the architecture. The central space is dominated both externally and internally by one of the largest domes in the world. The entrance is through a narthex, or entrance hall, which stretches across the building. One of the decorated bronze doors leading from the narthex is the Holy Door, only opened during jubilees.
The interior dimensions are vast when compared to other churches. One author wrote: "Only gradually does it dawn upon us – as we watch people draw near to this or that monument, strangely they appear to shrink; they are, of course, dwarfed by the scale of everything in the building. This in its turn overwhelms us."
The nave which leads to the central dome is in three bays, with piers supporting a barrel vault, the highest of any church. The nave is framed by wide aisles which have a number of chapels off them. There are also chapels surrounding the dome. Moving around the basilica in a clockwise direction they are: The Baptistery, the Chapel of the Presentation of the Virgin, the larger Choir Chapel, the altar of the Transfiguration, the Clementine Chapel with the altar of Saint Gregory, the Sacristy Entrance, the Altar of the Lie, the left transept with altars to the Crucifixion of Saint Peter, Saint Joseph and Saint Thomas, the altar of the Sacred Heart, the Chapel of the Madonna of Column, the altar of Saint Peter and the Paralytic, the apse with the Chair of Saint Peter, the altar of Saint Peter raising Tabitha, the altar of St. Petronilla, the altar of the Archangel Michael, the altar of the Navicella, the right transept with altars of Saint Erasmus, Saints Processo and Martiniano, and Saint Wenceslas, the altar of St. Jerome, the altar of Saint Basil, the Gregorian Chapel with the altar of the Madonna of Succour, the larger Chapel of the Holy Sacrament, the Chapel of Saint Sebastian and the Chapel of the Pietà. The Monuments, in a clockwise direction, are to: Maria Clementina Sobieski, The Stuarts, Benedict XV, John XXIII, St. Pius X, Innocent VIII, Leo XI, Innocent XI, Pius VII, Pius VIII, Alexander VII, Alexander VIII, Paul III, Urban VIII, Clement X, Clement XIII, Benedict XIV, St Peter (Bronze Statue), Gregory XVI, Gregory XIV, Gregory XIII, Matilda of Canossa, Innocent XII, Pius XII, Pius XI, Christina of Sweden, Leo XII.
At the heart of the basilica, beneath the high altar, is the Confessio or Chapel of the Confession, in reference to the confession of faith by St. Peter, which led to his martyrdom. Two curving marble staircases lead to this underground chapel at the level of the Constantinian church and immediately above the purported burial place of Saint Peter.
The entire interior of St. Peter's is lavishly decorated with marble, reliefs, architectural sculpture and gilding. The basilica contains a large number of tombs of popes and other notable people, many of which are considered outstanding artworks. There are also a number of sculptures in niches and chapels, including Michelangelo's Pietà. The central feature is a baldachin, or canopy over the Papal Altar, designed by Gian Lorenzo Bernini. The apse culminates in a sculptural ensemble, also by Bernini, and containing the symbolic Chair of Saint Peter.
One observer wrote: "St Peter's Basilica is the reason why Rome is still the centre of the civilized world. For religious, historical, and architectural reasons it by itself justifies a journey to Rome, and its interior offers a palimpsest of artistic styles at their best ..."
The American philosopher Ralph Waldo Emerson described St. Peter's as "an ornament of the earth ... the sublime of the beautiful."
Status
St. Peter's Basilica is one of the papal basilicas (previously styled "patriarchal basilicas") and one of the four Major Basilicas of Rome, the other Major Basilicas (all of which are also Papal Basilicas) being the Basilicas of St. John Lateran, St. Mary Major, and St. Paul outside the Walls. The rank of major basilica confers on St. Peter's Basilica precedence before all minor basilicas worldwide. However, unlike all the other Papal Major Basilicas, it is wholly within the territory, and thus the sovereign jurisdiction, of the Vatican City State, and not that of Italy. This is in contrast to the other three Papal Major Basilicas, which are within Italian territory and not the territory of the Vatican City State. (Lateran Treaty of 1929, Article 15 (Ibidem)) However, the Holy See fully owns these three basilicas, and Italy is legally obligated to recognize its full ownership thereof (Lateran Treaty of 1929, Article 13 (Ibidem)) and to concede to all of them "the immunity granted by International Law to the headquarters of the diplomatic agents of foreign States" (Lateran Treaty of 1929, Article 15 (Ibidem)).
It is the most prominent building in the Vatican City. Its dome is a feature of the skyline of Rome. Probably the largest church in Christendom, it covers an area of . One of the holiest sites of Christianity and Catholic Tradition, it is traditionally the burial site of its titular, St. Peter, who was the head of the twelve Apostles of Jesus and, according to tradition, the first Bishop of Antioch and later the first Bishop of Rome, rendering him the first Pope. Although the New Testament does not mention St. Peter's martyrdom in Rome, tradition, based on the writings of the Fathers of the Church, holds that his tomb is below the baldachin and the altar of the Basilica in the "Confession". For this reason, many Popes have, from the early years of the Church, been buried near Pope St. Peter in the necropolis beneath the Basilica. Construction of the current basilica, over the old Constantinian basilica, began on 18 April 1506 and finished in 1615. At length, on 18 November 1626 Pope Urban VIII solemnly dedicated the Basilica.
St. Peter's Basilica is neither the Pope's official seat nor first in rank among the Major Basilicas of Rome. This honour is held by the Pope's cathedral, the Archbasilica of St. John Lateran which is the mother church of all churches in communion with the Catholic Church. However, St. Peter's is certainly the Pope's principal church in terms of use because most Papal liturgies and ceremonies take place there due to its size, proximity to the Papal residence, and location within the Vatican City proper. The "Chair of Saint Peter", or cathedra, an ancient chair sometimes presumed to have been used by St. Peter himself, but which was a gift from Charles the Bald and used by many popes, symbolizes the continuing line of apostolic succession from St. Peter to the reigning Pope. It occupies an elevated position in the apse of the Basilica, supported symbolically by the Doctors of the Church and enlightened symbolically by the Holy Spirit.
As one of the constituent structures of the historically and architecturally significant Vatican City, St. Peter's Basilica was inscribed as a UNESCO World Heritage Site in 1984 under criteria (i), (ii), (iv), and (vi). With an exterior area of , an interior area of , St. Peter's Basilica is the largest Christian church building in the world by the two latter metrics and the second largest by the first . The top of its dome, at , also places it as the second tallest building in Rome . The dome's soaring height placed it among the tallest buildings of the Old World, and it continues to hold the title of tallest dome in the world. Though the largest dome in the world by diameter at the time of its completion, it no longer holds this distinction.
History
Saint Peter's burial site
After the crucifixion of Jesus, it is recorded in the Biblical book of the Acts of the Apostles that one of his twelve disciples, Simon known as Saint Peter, a fisherman from Galilee, took a leadership position among Jesus' followers and was of great importance in the founding of the Christian Church. The name Peter is "Petrus" in Latin and "Petros" in Greek, deriving from "petra" which means "stone" or "rock" in Greek, and is the literal translation of the Aramaic "Kepa", the name given to Simon by Jesus. (John 1:42, and see Matthew 16:18)
Catholic tradition holds that Peter, after a ministry of thirty-four years, travelled to Rome and met his martyrdom there along with Paul on 13 October 64 AD during the reign of the Roman Emperor Nero. His execution was one of the many martyrdoms of Christians following the Great Fire of Rome. According to Jerome, Peter was crucified head downwards, by his own request because he considered himself unworthy to die in the same manner as Jesus. The crucifixion took place near an ancient Egyptian obelisk in the Circus of Nero. The obelisk now stands in St. Peter's Square and is revered as a "witness" to Peter's death. It is one of several ancient Obelisks of Rome.
According to tradition, Peter's remains were buried just outside the Circus, on the Mons Vaticanus across the Via Cornelia from the Circus, less than from his place of death. The Via Cornelia was a road which ran east-to-west along the north wall of the Circus on land now covered by the southern portions of the Basilica and St. Peter's Square. A shrine was built on this site some years later. Almost three hundred years later, Old St. Peter's Basilica was constructed over this site.
The area now covered by the Vatican City had been a cemetery for some years before the Circus of Nero was built. It was a burial ground for the numerous executions in the Circus and contained many Christian burials because for many years after the burial of Saint Peter many Christians chose to be buried near Peter.
In 1939, in the reign of Pope Pius XII, 10 years of archaeological research began under the crypt of the basilica in an area inaccessible since the 9th century. The excavations revealed the remains of shrines of different periods at different levels, from Clement VIII (1594) to Callixtus II (1123) and Gregory I (590–604), built over an aedicula containing fragments of bones that were folded in a tissue with gold decorations, tinted with the precious murex purple. Although it could not be determined with certainty that the bones were those of Peter, the rare vestments suggested a burial of great importance. On 23 December 1950, in his pre-Christmas radio broadcast to the world, Pope Pius XII announced the discovery of Saint Peter's tomb.
Old St. Peter's Basilica
Old St. Peter's Basilica was the 4th-century church begun by the Emperor Constantine the Great between 319 and 333 AD. It was of typical basilical form, a wide nave and two aisles on each side and an apsidal end, with the addition of a transept or bema, giving the building the shape of a tau cross. It was over long, and the entrance was preceded by a large colonnaded atrium. This church had been built over the small shrine believed to mark the burial place of St. Peter, though the tomb was "smashed" in 846 AD. It contained a very large number of burials and memorials, including those of most of the popes from St. Peter to the 15th century. Like all of the earliest churches in Rome, both this church and its successor had the entrance to the east and the apse at the west end of the building. Since the construction of the current basilica, the name Old St. Peter's Basilica has been used for its predecessor to distinguish the two buildings.
Plan to rebuild
By the end of the 15th century, having been neglected during the period of the Avignon Papacy, the old basilica had fallen into disrepair. It appears that the first pope to consider rebuilding or at least making radical changes was Pope Nicholas V (1447–1455). He commissioned work on the old building from Leone Battista Alberti and Bernardo Rossellino and also had Rossellino design a plan for an entirely new basilica, or an extreme modification of the old. His reign was frustrated by political problems and when he died, little had been achieved. He had, however, ordered the demolition of the Colosseum and by the time of his death, 2,522 cartloads of stone had been transported for use in the new building. The foundations were completed for a new transept and choir to form a domed Latin cross with the preserved nave and side aisles of the old basilica. Some walls for the choir had also been built.
Pope Julius II planned far more for St Peter's than Nicholas V's program of repair or modification. Julius was at that time planning his own tomb, which was to be designed and adorned with sculpture by Michelangelo and placed within St Peter's. In 1505 Julius made a decision to demolish the ancient basilica and replace it with a monumental structure to house his enormous tomb and "aggrandize himself in the popular imagination". A competition was held, and a number of the designs have survived at the Uffizi Gallery. A succession of popes and architects followed in the next 120 years, their combined efforts resulting in the present building. The scheme begun by Julius II continued through the reigns of Leo X (1513–1521), Hadrian VI (1522–1523). Clement VII (1523–1534), Paul III (1534–1549), Julius III (1550–1555), Marcellus II (1555), Paul IV (1555–1559), Pius IV (1559–1565), Pius V (saint) (1565–1572), Gregory XIII (1572–1585), Sixtus V (1585–1590), Urban VII (1590), Gregory XIV (1590–1591), Innocent IX (1591), Clement VIII (1592–1605), Leo XI (1605), Paul V (1605–1621), Gregory XV (1621–1623), Urban VIII (1623–1644) and Innocent X (1644–1655).
Financing with indulgences
One method employed to finance the building of St. Peter's Basilica was the granting of indulgences in return for contributions. A major promoter of this method of fund-raising was Albrecht, Archbishop of Mainz and Magdeburg, who had to clear debts owed to the Roman Curia by contributing to the rebuilding program. To facilitate this, he appointed the German Dominican preacher Johann Tetzel, whose salesmanship provoked a scandal.
A German Augustinian priest, Martin Luther, wrote to Archbishop Albrecht arguing against this "selling of indulgences". He also included his "Disputation of Martin Luther on the Power and Efficacy of Indulgences", which came to be known as The 95 Theses. This became a factor in starting the Reformation, the birth of Protestantism.
Architecture
Successive plans
Pope Julius' scheme for the grandest building in Christendom was the subject of a competition for which a number of entries remain intact in the Uffizi Gallery, Florence. It was the design of Donato Bramante that was selected, and for which the foundation stone was laid in 1506. This plan was in the form of an enormous Greek Cross with a dome inspired by that of the huge circular Roman temple, the Pantheon. The main difference between Bramante's design and that of the Pantheon is that where the dome of the Pantheon is supported by a continuous wall, that of the new basilica was to be supported only on four large piers. This feature was maintained in the ultimate design. Bramante's dome was to be surmounted by a lantern with its own small dome but otherwise very similar in form to the Early Renaissance lantern of Florence Cathedral designed for Brunelleschi's dome by Michelozzo.
Bramante had envisioned that the central dome would be surrounded by four lower domes at the diagonal axes. The equal chancel, nave and transept arms were each to be of two bays ending in an apse. At each corner of the building was to stand a tower, so that the overall plan was square, with the apses projecting at the cardinal points. Each apse had two large radial buttresses, which squared off its semi-circular shape.
When Pope Julius died in 1513, Bramante was replaced with Giuliano da Sangallo and Fra Giocondo, who both died in 1515 (Bramante himself having died the previous year). Raphael was confirmed as the architect of St. Peter's on 1 August 1514. The main change in his plan is the nave of five bays, with a row of complex apsidal chapels off the aisles on either side. Raphael's plan for the chancel and transepts made the squareness of the exterior walls more definite by reducing the size of the towers, and the semi-circular apses more clearly defined by encircling each with an ambulatory.
In 1520 Raphael also died, aged 37, and his successor Baldassare Peruzzi maintained changes that Raphael had proposed to the internal arrangement of the three main apses, but otherwise reverted to the Greek Cross plan and other features of Bramante. This plan did not go ahead because of various difficulties of both Church and state. In 1527 Rome was sacked and plundered by Emperor Charles V. Peruzzi died in 1536 without his plan being realized.
At this point Antonio da Sangallo the Younger submitted a plan which combines features of Peruzzi, Raphael and Bramante in its design and extends the building into a short nave with a wide façade and portico of dynamic projection. His proposal for the dome was much more elaborate of both structure and decoration than that of Bramante and included ribs on the exterior. Like Bramante, Sangallo proposed that the dome be surmounted by a lantern which he redesigned to a larger and much more elaborate form. Sangallo's main practical contribution was to strengthen Bramante's piers which had begun to crack.
On 1 January 1547 in the reign of Pope Paul III, Michelangelo, then in his seventies, succeeded Sangallo the Younger as "Capomaestro", the superintendent of the building program at St Peter's. He is to be regarded as the principal designer of a large part of the building as it stands today, and as bringing the construction to a point where it could be carried through. He did not take on the job with pleasure; it was forced upon him by Pope Paul, frustrated at the death of his chosen candidate, Giulio Romano and the refusal of Jacopo Sansovino to leave Venice. Michelangelo wrote, "I undertake this only for the love of God and in honour of the Apostle." He insisted that he should be given a free hand to achieve the ultimate aim by whatever means he saw fit.
Michelangelo's contribution
Michelangelo took over a building site at which four piers, enormous beyond any constructed since ancient Roman times, were rising behind the remaining nave of the old basilica. He also inherited the numerous schemes designed and redesigned by some of the greatest architectural and engineering minds of the 16th century. There were certain common elements in these schemes. They all called for a dome to equal that engineered by Brunelleschi a century earlier and which has since dominated the skyline of Renaissance Florence, and they all called for a strongly symmetrical plan of either Greek Cross form, like the iconic St. Mark's Basilica in Venice, or of a Latin Cross with the transepts of identical form to the chancel, as at Florence Cathedral.
Even though the work had progressed only a little in 40 years, Michelangelo did not simply dismiss the ideas of the previous architects. He drew on them in developing a grand vision. Above all, Michelangelo recognized the essential quality of Bramante's original design. He reverted to the Greek Cross and, as Helen Gardner expresses it: "Without destroying the centralising features of Bramante's plan, Michelangelo, with a few strokes of the pen converted its snowflake complexity into massive, cohesive unity."
As it stands today, St. Peter's has been extended with a nave by Carlo Maderno. It is the chancel end (the ecclesiastical "Eastern end") with its huge centrally placed dome that is the work of Michelangelo. Because of its location within the Vatican State and because the projection of the nave screens the dome from sight when the building is approached from the square in front of it, the work of Michelangelo is best appreciated from a distance. What becomes apparent is that the architect has greatly reduced the clearly defined geometric forms of Bramante's plan of a square with square projections, and also of Raphael's plan of a square with semi-circular projections. Michelangelo has blurred the definition of the geometry by making the external masonry of massive proportions and filling in every corner with a small vestry or stairwell. The effect created is of a continuous wall surface that is folded or fractured at different angles, but lacks the right angles which usually define change of direction at the corners of a building. This exterior is surrounded by a giant order of Corinthian pilasters all set at slightly different angles to each other, in keeping with the ever-changing angles of the wall's surface. Above them, the huge cornice ripples in a continuous band, giving the appearance of keeping the whole building in a state of compression.
Dome: successive and final designs
The dome of St. Peter's rises to a total height of from the floor of the basilica to the top of the external cross. It is the tallest dome in the world. Its internal diameter is , slightly smaller than two of the three other huge domes that preceded it, those of the Pantheon of Ancient Rome, , and Florence Cathedral of the Early Renaissance, . It has a greater diameter by approximately than Constantinople's Hagia Sophia church, completed in 537. It was to the domes of the Pantheon and Florence duomo that the architects of St. Peter's looked for solutions as to how to go about building what was conceived, from the outset, as the greatest dome of Christendom.
Bramante and Sangallo, 1506 and 1513
The dome of the Pantheon stands on a circular wall with no entrances or windows except a single door. The whole building is as high as it is wide. Its dome is constructed in a single shell of concrete, made light by the inclusion of a large amount of the volcanic stones tuff and pumice. The inner surface of the dome is deeply coffered which has the effect of creating both vertical and horizontal ribs while lightening the overall load. At the summit is an ocular opening across which provides light to the interior.
Bramante's plan for the dome of St. Peter's (1506) follows that of the Pantheon very closely, and like that of the Pantheon, was designed to be constructed in Tufa Concrete for which he had rediscovered a formula. With the exception of the lantern that surmounts it, the profile is very similar, except that in this case, the supporting wall becomes a drum raised high above ground level on four massive piers. The solid wall, as used at the Pantheon, is lightened at St. Peter's by Bramante piercing it with windows and encircling it with a peristyle.
In the case of Florence Cathedral, the desired visual appearance of the pointed dome existed for many years before Brunelleschi made its construction feasible. Its double-shell construction of bricks locked together in a herringbone pattern (re-introduced from Byzantine architecture), and the gentle upward slope of its eight stone ribs made it possible for the construction to take place without the massive wooden formwork necessary to construct hemispherical arches. While its appearance, with the exception of the details of the lantern, is entirely Gothic, its engineering was highly innovative, and the product of a mind that had studied the huge vaults and remaining dome of Ancient Rome.
Sangallo's plan (1513), of which a large wooden model still exists, looks to both these predecessors. He realized the value of both the coffering at the Pantheon and the outer stone ribs at Florence Cathedral. He strengthened and extended the peristyle of Bramante into a series of arched and ordered openings around the base, with a second such arcade set back in a tier above the first. In his hands, the rather delicate form of the lantern, based closely on that in Florence, became a massive structure, surrounded by a projecting base, a peristyle and surmounted by a spire of conic form. According to James Lees-Milne the design was "too eclectic, too pernickety and too tasteless to have been a success".
Michelangelo and Giacomo della Porta, 1547 and 1585
Michelangelo redesigned the dome in 1547, taking into account all that had gone before. His dome, like that of Florence, is constructed of two shells of brick, the outer one having 16 stone ribs, twice the number at Florence but far fewer than in Sangallo's design. As with the designs of Bramante and Sangallo, the dome is raised from the piers on a drum. The encircling peristyle of Bramante and the arcade of Sangallo are reduced to 16 pairs of Corinthian columns, each of high which stand proud of the building, connected by an arch. Visually they appear to buttress each of the ribs, but structurally they are probably quite redundant. The reason for this is that the dome is ovoid in shape, rising steeply as does the dome of Florence Cathedral, and therefore exerting less outward thrust than does a hemispherical dome, such as that of the Pantheon, which, although it is not buttressed, is countered by the downward thrust of heavy masonry which extends above the circling wall.
The ovoid profile of the dome has been the subject of much speculation and scholarship over the past century. Michelangelo died in 1564, leaving the drum of the dome complete, and Bramante's piers much bulkier than originally designed, each across. Following his death, the work continued under his assistant Jacopo Barozzi da Vignola with Giorgio Vasari appointed by Pope Pius V as a watchdog to make sure that Michelangelo's plans were carried out exactly. Despite Vignola's knowledge of Michelangelo's intentions, little happened in this period. In 1585 the energetic Pope Sixtus V appointed Giacomo della Porta who was to be assisted by Domenico Fontana. The five-year reign of Sixtus was to see the building advance at a great rate.
Michelangelo left a few drawings, including an early drawing of the dome, and some details. There were also detailed engravings published in 1569 by Stefan du Pérac who claimed that they were the master's final solution. Michelangelo, like Sangallo before him, also left a large wooden model. Giacomo della Porta subsequently altered this model in several ways. The major change restored an earlier design, in which the outer dome appears to rise above, rather than rest directly on the base. Most of the other changes were of a cosmetic nature, such as the adding of lion's masks over the swags on the drum in honour of Pope Sixtus and adding a circlet of finials around the spire at the top of the lantern, as proposed by Sangallo.
A drawing by Michelangelo indicates that his early intentions were towards an ovoid dome, rather than a hemispherical one. In an engraving in Galasso Alghisi' treatise (1563), the dome may be represented as ovoid, but the perspective is ambiguous. Stefan du Pérac's engraving (1569) shows a hemispherical dome, but this was perhaps an inaccuracy of the engraver. The profile of the wooden model is more ovoid than that of the engravings, but less so than the finished product. It has been suggested that Michelangelo on his death bed reverted to the more pointed shape. However, Lees-Milne cites Giacomo della Porta as taking full responsibility for the change and as indicating to Pope Sixtus that Michelangelo was lacking in the scientific understanding of which he himself was capable.
Helen Gardner suggests that Michelangelo made the change to the hemispherical dome of lower profile in order to establish a balance between the dynamic vertical elements of the encircling giant order of pilasters and a more static and reposeful dome. Gardner also comments, "The sculpturing of architecture [by Michelangelo] ... here extends itself up from the ground through the attic stories and moves on into the drum and dome, the whole building being pulled together into a unity from base to summit."
It is this sense of the building being sculptured, unified and "pulled together" by the encircling band of the deep cornice that led Eneide Mignacca to conclude that the ovoid profile, seen now in the end product, was an essential part of Michelangelo's first (and last) concept. The sculptor/architect has, figuratively speaking, taken all the previous designs in hand and compressed their contours as if the building were a lump of clay. The dome must appear to thrust upwards because of the apparent pressure created by flattening the building's angles and restraining its projections. If this explanation is the correct one, then the profile of the dome is not merely a structural solution, as perceived by Giacomo della Porta; it is part of the integrated design solution that is about visual tension and compression. In one sense, Michelangelo's dome may appear to look backward to the Gothic profile of Florence Cathedral and ignore the Classicism of the Renaissance, but on the other hand, perhaps more than any other building of the 16th century, it prefigures the architecture of the Baroque.
Completion
Giacomo della Porta and Domenico Fontana brought the dome to completion in 1590, the last year of the reign of Sixtus V. His successor, Gregory XIV, saw Fontana complete the lantern and had an inscription to the honour of Sixtus V placed around its inner opening. The next pope, Clement VIII, had the cross raised into place, an event which took all day, and was accompanied by the ringing of the bells of all the city's churches. In the arms of the cross are set two lead caskets, one containing a fragment of the True Cross and a relic of St. Andrew and the other containing medallions of the Holy Lamb.
In the mid-18th century, cracks appeared in the dome, so four iron chains were installed between the two shells to bind it, like the rings that keep a barrel from bursting. As many as ten chains have been installed at various times, the earliest possibly planned by Michelangelo himself as a precaution, as Brunelleschi did at Florence Cathedral.
Around the inside of the dome is written, in letters high:
Beneath the lantern is the inscription:
Discovery of Michelangelo draft
On 7 December 2007, a fragment of a red chalk drawing of a section of the dome of the basilica, almost certainly by the hand of Michelangelo, was discovered in the Vatican archives. The drawing shows a small precisely drafted section of the plan of the entablature above two of the radial columns of the cupola drum. Michelangelo is known to have destroyed thousands of his drawings before his death. The rare survival of this example is probably due to its fragmentary state and the fact that detailed mathematical calculations had been made over the top of the drawing.
Addition of nave and facade
On 18 February 1606, under Pope Paul V, the dismantling of the remaining parts of the Constantinian basilica began. The marble cross that had been set at the top of the pediment by Pope Sylvester and Constantine the Great was lowered to the ground. The timbers were salvaged for the roof of the Borghese Palace and two rare black marble columns, the largest of their kind, were carefully stored and later used in the narthex. The tombs of various popes were opened, treasures removed and plans made for re-interment in the new basilica.
The Pope had appointed Carlo Maderno in 1602. He was a nephew of Domenico Fontana and had demonstrated himself as a dynamic architect. Maderno's idea was to ring Michelangelo's building with chapels, but the Pope was hesitant about deviating from the master's plan, even though he had been dead for forty years. The Fabbrica or building committee, a group drawn from various nationalities and generally despised by the Curia who viewed the basilica as belonging to Rome rather than Christendom, were in a quandary as to how the building should proceed. One of the matters that influenced their thinking was the Counter-Reformation which increasingly associated a Greek Cross plan with paganism and saw the Latin Cross as truly symbolic of Christianity. The central plan also did not have a "dominant orientation toward the east."
Another influence on the thinking of both the Fabbrica and the Curia was a certain guilt at the demolition of the ancient building. The ground on which it and its various associated chapels, vestries and sacristies had stood for so long was hallowed. The only solution was to build a nave that encompassed the whole space. In 1607 a committee of ten architects was called together, and a decision was made to extend Michelangelo's building into a nave. Maderno's plans for both the nave and the facade were accepted.
The building of the nave began on 7 May 1607, and proceeded at a great rate, with an army of 700 labourers being employed. The following year, the façade was begun, in December 1614 the final touches were added to the stucco decoration of the vault and early in 1615 the partition wall between the two sections was pulled down. All the rubble was carted away, and the nave was ready for use by Palm Sunday.
Maderno's facade
The facade designed by Maderno, is wide and high and is built of travertine stone, with a giant order of Corinthian columns and a central pediment rising in front of a tall attic surmounted by thirteen statues: Christ flanked by eleven of the Apostles (except Saint Peter, whose statue is left of the stairs) and John the Baptist. The inscription below the cornice on the tall frieze reads:
(Paul V (Camillo Borghese), born in Rome but of a Sienese family, liked to emphasize his "Romanness.")
The facade is often cited as the least satisfactory part of the design of St. Peter's. The reasons for this, according to James Lees-Milne, are that it was not given enough consideration by the Pope and committee because of the desire to get the building completed quickly, coupled with the fact that Maderno was hesitant to deviate from the pattern set by Michelangelo at the other end of the building. Lees-Milne describes the problems of the façade as being too broad for its height, too cramped in its details and too heavy in the attic story. The breadth is caused by modifying the plan to have towers on either side. These towers were never executed above the line of the facade because it was discovered that the ground was not sufficiently stable to bear the weight. One effect of the facade and lengthened nave is to screen the view of the dome, so that the building, from the front, has no vertical feature, except from a distance.
Narthex and portals
Behind the façade of St. Peter's stretches a long portico or "narthex" such as was occasionally found in Italian churches. This is the part of Maderno's design with which he was most satisfied. Its long barrel vault is decorated with ornate stucco and gilt, and successfully illuminated by small windows between pendentives, while the ornate marble floor is beamed with light reflected in from the piazza. At each end of the narthex is a theatrical space framed by ionic columns and within each is set a statue, an equestrian statue of Charlemagne (18th century) by Cornacchini in the south end and The Vision of Constantine (1670) by Bernini in the north end.
Five portals, of which three are framed by huge salvaged antique columns, lead into the basilica. The central portal has a bronze door created by Antonio Averulino c. 1440 for the old basilica and somewhat enlarged to fit the new space.
Maderno's nave
To the single bay of Michelangelo's Greek Cross, Maderno added a further three bays. He made the dimensions slightly different from Michelangelo's bay, thus defining where the two architectural works meet. Maderno also tilted the axis of the nave slightly. This was not by accident, as suggested by his critics. An ancient Egyptian obelisk had been erected in the square outside, but had not been quite aligned with Michelangelo's building, so Maderno compensated, in order that it should, at least, align with the Basilica's façade.
The nave has huge paired pilasters, in keeping with Michelangelo's work. The size of the interior is so "stupendously large" that it is hard to get a sense of scale within the building. The four cherubs who flutter against the first piers of the nave, carrying between them two holy water basins, appear of quite normal cherubic size, until approached. Then it becomes apparent that each one is over 2 metres high and that real children cannot reach the basins unless they scramble up the marble draperies. The aisles each have two smaller chapels and a larger rectangular chapel, the Chapel of the Sacrament and the Choir Chapel. These are lavishly decorated with marble, stucco, gilt, sculpture and mosaic. Remarkably, all of the large altarpieces, with the exception of the Holy Trinity by Pietro da Cortona in the Blessed Sacrament Chapel, have been reproduced in mosaic. Two precious paintings from the old basilica, Our Lady of Perpetual Help and Our Lady of the Column are still being used as altarpieces.
Maderno's last work at St. Peter's was to design a crypt-like space or "Confessio" under the dome, where the cardinals and other privileged persons could descend in order to be nearer to the burial place of the apostle. Its marble steps are remnants of the old basilica and around its balustrade are 95 bronze lamps.
Influence on church architecture
The design of St. Peter's Basilica, and in particular its dome, has greatly influenced church architecture in Western Christendom. Within Rome, the huge domed church of Sant'Andrea della Valle was designed by Giacomo della Porta before the completion of St Peter's Basilica, and subsequently worked on by Carlo Maderno. This was followed by the domes of San Carlo ai Catinari, Sant'Agnese in Agone, and many others. Christopher Wren's dome at St Paul's Cathedral (London, England), the domes of Karlskirche (Vienna, Austria), St. Nicholas Church (Prague, Czech Republic), and the Pantheon (Paris, France) all pay homage to St Peter's Basilica.
The 19th and early-20th-century architectural revivals brought about the building of a great number of churches that imitate elements of St Peter's to a greater or lesser degree, including St. Mary of the Angels in Chicago, St. Josaphat's Basilica in Milwaukee, Immaculate Heart of Mary in Pittsburgh and Mary, Queen of the World Cathedral in Montreal, which replicates many aspects of St Peter's on a smaller scale. Post-Modernism has seen free adaptations of St Peter's in the Basilica of Our Lady of Licheń, and the Basilica of Our Lady of Peace of Yamoussoukro.
Bernini's furnishings
Pope Urban VIII and Bernini
As a young boy Gian Lorenzo Bernini (1598–1680) visited St. Peter's with the painter Annibale Carracci and stated his wish to build "a mighty throne for the apostle". His wish came true. As a young man, in 1626, he received the patronage of Pope Urban VIII and worked on the embellishment of the Basilica for 50 years. Appointed as Maderno's successor in 1629, he was to become regarded as the greatest architect and sculptor of the Baroque period. Bernini's works at St. Peter's include the baldachin (baldaquin, from Italian: baldacchino), the Chapel of the Sacrament, the plan for the niches and loggias in the piers of the dome and the chair of St. Peter.
Baldacchino and niches
Bernini's first work at St. Peter's was to design the baldacchino, a pavilion-like structure tall and claimed to be the largest piece of bronze in the world, which stands beneath the dome and above the altar. Its design is based on the ciborium, of which there are many in the churches of Rome, serving to create a sort of holy space above and around the table on which the Sacrament is laid for the Eucharist and emphasizing the significance of this ritual. These ciboria are generally of white marble, with inlaid coloured stone. Bernini's concept was for something very different. He took his inspiration in part from the baldachin or canopy carried above the head of the pope in processions, and in part from eight ancient columns that had formed part of a screen in the old basilica. Their twisted barley-sugar shape had a special significance as they were modelled on those of the Temple of Jerusalem and donated by the Emperor Constantine. Based on these columns, Bernini created four huge columns of bronze, twisted and decorated with laurel leaves and bees, which were the emblem of Pope Urban.
The baldacchino is surmounted not with an architectural pediment, like most baldacchini, but with curved Baroque brackets supporting a draped canopy, like the brocade canopies carried in processions above precious iconic images. In this case, the draped canopy is of bronze, and all the details, including the olive leaves, bees, and the portrait heads of Urban's niece in childbirth and her newborn son, are picked out in gold leaf. The baldacchino stands as a vast free-standing sculptural object, central to and framed by the largest space within the building. It is so large that the visual effect is to create a link between the enormous dome which appears to float above it, and the congregation at floor level of the basilica. It is penetrated visually from every direction, and is visually linked to the Cathedra Petri in the apse behind it and to the four piers containing large statues that are at each diagonal.
As part of the scheme for the central space of the church, Bernini had the huge piers, begun by Bramante and completed by Michelangelo, hollowed out into niches, and had staircases made inside them, leading to four balconies. There was much dismay from those who thought that the dome might fall, but it did not. On the balconies Bernini created showcases, framed by the eight ancient twisted columns, to display the four most precious relics of the basilica: the spear of Longinus, said to have pierced the side of Christ, the veil of Veronica, with the miraculous image of the face of Christ, a fragment of the True Cross discovered in Jerusalem by Constantine's mother, Helena, and a relic of Saint Andrew, the brother of Saint Peter. In each of the niches that surround the central space of the basilica was placed a huge statue of the saint associated with the relic above. Only Saint Longinus is the work of Bernini. (See below)
Bernini's Towers
Urban had long been a critic of Bernini's predecessor, Carlo Maderno. His disapproval of the architect's work stemmed largely from the Maderno's design for the longitudinal nave of St. Peters, which was widely condemned for obscuring Michelangelo's dome. When the Pope gave the commission to Bernini he therefore requested that a new design for the facade's bell towers to be submitted for consideration. Baldinucci describes Bernini's tower as consisting of "two orders of columns and pilasters, the first order being Corinthian" and "a third or attic story formed of pilasters and two columns on either side of the open archway in the center".
Urban desired the towers to be completed by a very specific date: 29 June 1641, the feast day dedicated to Saints Peter and Paul. To this end an order was issued which stated that "all work should take a second seat to that of the campanile." The south tower was completed on time even in spite of these issues, but records show that in the wake of the unveiling the Pope was not content with what he saw and he ordered the top level of Bernini's tower removed so that the structure could be made even grander. The tower continued to grow, and as the construction began to settle, the first cracks started to appear followed by Urban's infamous public admonishment of his architect.
In 1642 all work on both towers came to a halt. Bernini had to pay the cost for the demolition; eventually the idea of completing the bell towers was abandoned.
Cathedra Petri and Chapel of the Blessed Sacrament
Bernini then turned his attention to another precious relic, the so-called Cathedra Petri or "throne of St. Peter" a chair which was often claimed to have been used by the apostle, but appears to date from the 12th century. As the chair itself was fast deteriorating and was no longer serviceable, Pope Alexander VII determined to enshrine it in suitable splendor as the object upon which the line of successors to Peter was based. Bernini created a large bronze throne in which it was housed, raised high on four looping supports held effortlessly by massive bronze statues of four Doctors of the Church, Saints Ambrose and Augustine representing the Latin Church and Athanasius and John Chrysostom, the Greek Church. The four figures are dynamic with sweeping robes and expressions of adoration and ecstasy. Behind and above the cathedra, a blaze of light comes in through a window of yellow alabaster, illuminating, at its centre, the Dove of the Holy Spirit. The elderly painter, Andrea Sacchi, had urged Bernini to make the figures large, so that they would be seen well from the central portal of the nave. The chair was enshrined in its new home with great celebration of 16 January 1666.
Bernini's final work for St. Peter's, undertaken in 1676, was the decoration of the Chapel of the Sacrament. To hold the sacramental Host, he designed a miniature version in gilt bronze of Bramante's Tempietto, the little chapel that marks the place of the death of St. Peter. On either side is an angel, one gazing in rapt adoration and the other looking towards the viewer in welcome. Bernini died in 1680 in his 82nd year.
St. Peter's Piazza
To the east of the basilica is the Piazza di San Pietro, (St. Peter's Square). The present arrangement, constructed between 1656 and 1667, is the Baroque inspiration of Bernini who inherited a location already occupied by an Egyptian obelisk which was centrally placed, (with some contrivance) to Maderno's facade. The obelisk, known as "The Witness", at and a total height, including base and the cross on top, of , is the second largest standing obelisk, and the only one to remain standing since its removal from Egypt and re-erection at the Circus of Nero in 37 AD, where it is thought to have stood witness to the crucifixion of Saint Peter. Its removal to its present location by order of Pope Sixtus V and engineered by Domenico Fontana on 28 September 1586, was an operation fraught with difficulties and nearly ending in disaster when the ropes holding the obelisk began to smoke from the friction. Fortunately this problem was noticed by Benedetto Bresca, a sailor of Sanremo, and for his swift intervention, his town was granted the privilege of providing the palms that are used at the basilica each Palm Sunday.
The other object in the old square with which Bernini had to contend was a large fountain designed by Maderno in 1613 and set to one side of the obelisk, making a line parallel with the facade. Bernini's plan uses this horizontal axis as a major feature of his unique, spatially dynamic and highly symbolic design. The most obvious solutions were either a rectangular piazza of vast proportions so that the obelisk stood centrally and the fountain (and a matching companion) could be included, or a trapezoid piazza which fanned out from the facade of the basilica like that in front of the Palazzo Pubblico in Siena. The problems of the square plan are that the necessary width to include the fountain would entail the demolition of numerous buildings, including some of the Vatican, and would minimize the effect of the facade. The trapezoid plan, on the other hand, would maximize the apparent width of the facade, which was already perceived as a fault of the design.
Bernini's ingenious solution was to create a piazza in two sections. That part which is nearest the basilica is trapezoid, but rather than fanning out from the facade, it narrows. This gives the effect of countering the visual perspective. It means that from the second part of the piazza, the building looks nearer than it is, the breadth of the facade is minimized and its height appears greater in proportion to its width. The second section of the piazza is a huge elliptical circus which gently slopes downwards to the obelisk at its centre. The two distinct areas are framed by a colonnade formed by doubled pairs of columns supporting an entablature of the simple Tuscan Order.
The part of the colonnade that is around the ellipse does not entirely encircle it, but reaches out in two arcs, symbolic of the arms of "the Catholic Church reaching out to welcome its communicants". The obelisk and Maderno's fountain mark the widest axis of the ellipse. Bernini balanced the scheme with another fountain in 1675. The approach to the square used to be through a jumble of old buildings, which added an element of surprise to the vista that opened up upon passing through the colonnade. Nowadays a long wide street, the Via della Conciliazione, built by Mussolini after the conclusion of the Lateran Treaties, leads from the River Tiber to the piazza and gives distant views of St. Peter's as the visitor approaches, with the basilica acting as a terminating vista.
Bernini's transformation of the site is entirely Baroque in concept. Where Bramante and Michelangelo conceived a building that stood in "self-sufficient isolation", Bernini made the whole complex "expansively relate to its environment". Banister Fletcher says "No other city has afforded such a wide-swept approach to its cathedral church, no other architect could have conceived a design of greater nobility ... (it is) the greatest of all atriums before the greatest of all churches of Christendom."
Clocks
The top of the facade of St. Peter's Basilica has two clocks and several sculptures. The clocks were created to replace Bernini's bell towers which had to be torn down due to insufficient support. The left clock shows Rome time, the one of the right shows European mean time. The statues are Christ the Redeemer, St. John the Baptist and 11 Apostles. From the left: St. Thadeus, St. Matthew, St. Philip, St. Thomas, St. James the Greater, St. John the Baptist, The Redeemer, St. Andrew, St. John the Evangelist, St. James the Lesser, St. Bartholomew, St. Simeon, and St. Matthias. Above the Roman clock is the coat of arms for the city-state of Vatican City since 1931 held by two angels.
Bells
The Basilica has 6 bells, placed in the room under the Roman clock, only 3 of them are visible from ground level while the rest are hidden behind the bourdon. They range from the smallest which is 260 kg to the massive bourdon that approximately weighs 9 tonnes. From 1931, the bells are operated electrically, thus permitting even the largest bell to be tolled from a distance. The oldest bell Rota dates from 1288 and the bourdon called Campanone is rung at Christmas and Easter, on the Solemnity of Sts. Peter and Paul, and every time the Pope imparts the "Urbi et Orbi" blessing to the city and to the world. Campanone also announces the election of a new pope.
Treasures
Tombs and relics
There are over 100 tombs within St. Peter's Basilica (extant to various extents), many located beneath the Basilica. These include 91 popes, Saint Ignatius of Antioch, Holy Roman Emperor Otto II, and the composer Giovanni Pierluigi da Palestrina. Exiled Catholic British royalty James Francis Edward Stuart and his two sons, Charles Edward Stuart and Henry Benedict Stuart, Cardinal Bishop of Frascati, are buried here, having been granted asylum by Pope Clement XI. Also buried here are Maria Clementina Sobieska, wife of James Francis Edward Stuart, Queen Christina of Sweden, who abdicated her throne in order to convert to Catholicism, and Countess Matilda of Tuscany, supporter of the Papacy during the Investiture Controversy. The most recent interment was Pope John Paul II, on 8 April 2005. Beneath, near the crypt, is the recently discovered vaulted 4th-century "Tomb of the Julii". (See below for some descriptions of tombs).
Artworks
Towers and narthex
In the towers to either side of the facade are two clocks. The clock on the left has been operated electrically since 1931. Its oldest bell dates from 1288.
One of the most important treasures of the basilica is a mosaic set above the central external door. Called the "Navicella", it is based on a design by Giotto (early 14th century) and represents a ship symbolizing the Christian Church. The mosaic is mostly a 17th-century copy of Giotto's original.
At each end of the narthex is an equestrian figure, to the north Constantine the Great by Bernini (1670) and to the south Charlemagne by Cornacchini (18th century).
Of the five portals from the narthex to the interior, three contain notable doors. The central portal has the Renaissance bronze door by Antonio Averulino (called Filarete) (1455), enlarged to fit the new space. The southern door, the Door of the Dead, was designed by 20th-century sculptor Giacomo Manzù and includes a portrait of Pope John XXIII kneeling before the crucified figure of Saint Peter.
The northernmost door is the "Holy Door" which, by tradition, is walled-up with bricks, and opened only for holy years such as the Jubilee year by the Pope. The present door is bronze and was designed by Vico Consorti in 1950 and cast in Florence by the Ferdinando Marinelli Artistic Foundry. Above it are inscriptions commemorating the opening of the door: PAVLVS V PONT MAX ANNO XIII and GREGORIVS XIII PONT MAX.
Recently installed commemorative plaques read above the door as follows:
Older commemorative plaques are removed to make way for the new plaque when the holy door is opened and sealed.
Nave
On the first piers of the nave are two Holy Water basins held by pairs of cherubs each 2 metres high, commissioned by Pope Benedict XIII from designer Agostino Cornacchini and sculptor Francesco Moderati, (1720s).
Along the floor of the nave are markers showing the comparative lengths of other churches, starting from the entrance.
On the decorative pilasters of the piers of the nave are medallions with relief depicting 56 of the first popes.
In niches between the pilasters of the nave are statues depicting 39 founders of religious orders.
Set against the north east pier of the dome is a statue of Saint Peter Enthroned, sometimes attributed to late 13th-century sculptor Arnolfo di Cambio, with some scholars dating it to the 5th century. One foot of the statue is largely worn away by pilgrims kissing it for centuries.
The sunken Confessio leading to the Vatican Grottoes (see above) contained a large kneeling statue by Canova of Pope Pius VI, who was captured and mistreated by Napoleon Bonaparte's army. This has now been moved to the back (eastern) end of the grottoes.
In the Confessio is the Niche of the Pallium ("Niche of Stoles") which contains a bronze urn, donated by Pope Benedict XIV, to contain white stoles embroidered with black crosses and woven with the wool of lambs blessed on St. Agnes' day.
The High Altar is surmounted by Bernini's baldachin. (See above)
Set in niches within the four piers supporting the dome are the larger-than-life statues associated with the basilica's primary holy relics: Saint Helena holding the True Cross and the Holy Nails, by Andrea Bolgi; Saint Longinus holding the spear that pierced the side of Jesus, by Bernini (1639); Saint Andrew with the St. Andrew's Cross, by Francois Duquesnoy and Saint Veronica holding her veil with the image of Jesus' face, by Francesco Mochi.
North aisle
In the first chapel of the north aisle is Michelangelo's Pietà.
On the first pier in the right aisle is the monument of Queen Christina of Sweden, who abdicated in 1654 in order to convert to Catholicism.
The second chapel, dedicated to Saint Sebastian, contains the statues of popes Pius XI and Pius XII. The space below the altar used to be the resting place of Pope Innocent XI but his remains were moved to the Altar of the Transfiguration on 8 April 2011. This was done to make way for the body of Pope John Paul II. His remains were placed beneath the altar on 2 May 2011.
The large chapel on the right aisle is the Chapel of the Blessed Sacrament which contains the tabernacle by Bernini (1664) resembling Bramante's Tempietto at San Pietro in Montorio supported by two kneeling angels and with behind it a painting of the Holy Trinity by Pietro da Cortona.
Near the altar of Our Lady of Succour are the monuments of popes Gregory XIII by Camillo Rusconi (1723) and Gregory XIV.
At the end of the aisle is an altar containing the relics of Saint Petronilla and with an altarpiece The Burial of St Petronilla by Guercino (Giovanni Francesco Barbieri), 1623.
South aisle
The first chapel in the south aisle is the baptistry, commissioned by Pope Innocent XII and designed by Carlo Fontana, (great nephew of Domenico Fontana). The font was carved from the lid of the purple porphyry sarcophagus which had once held the remains of the Emperor Hadrian, and is surmounted with a gilt-bronze figure of the "Lamb of God". Fontana's reworked porphyry sarcophagus lid font replaced an earlier font, which was re-purposed from the sarcophagus of Probus, the 4th-century Prefect of Rome, and which was used for baptisms from the 15th century until the late 17th century, when Fontana's work was completed.
Against the first pier of the aisle is the Monument to the Royal Stuarts: James Stuart, known as the "Old Pretender" and his sons, Charles Edward, known as "Bonnie Prince Charlie", and Henry, a cardinal. The tomb is a Neo-Classical design by Canova unveiled in 1819. Opposite it is the memorial of James Francis Edward Stuart's wife, Maria Clementina Sobieska.
The second chapel is that of the Presentation of the Virgin and contains the memorials of Pope Benedict XV and Pope John XXIII.
Against the piers are the tombs of Pope Pius X and Pope Innocent VIII.
The large chapel off the south aisle is the Choir Chapel which contains the altar of the Immaculate Conception.
At the entrance to the Sacristy is the tomb of Pope Pius VIII
The south transept contains the altars of Saint Thomas, Saint Joseph and the Crucifixion of Saint Peter.
The tomb of Fabio Chigi, Pope Alexander VII, towards the end of the aisle, is the work of Bernini and called by Lees-Milne "one of the greatest tombs of the Baroque Age". It occupies an awkward position, being set in a niche above a doorway into a small vestry, but Bernini has utilized the doorway in a symbolic manner. Pope Alexander kneels upon his tomb, facing outward. The tomb is supported on a large draped shroud in patterned red marble, and is supported by four female figures, of whom only the two at the front are fully visible. They represent Charity and Truth. The foot of Truth rests upon a globe of the world, her toe being pierced symbolically by the thorn of Protestant England. Coming forth, seemingly, from the doorway as if it were the entrance to a tomb, is the skeletal winged figure of Death, its head hidden beneath the shroud, but its right hand carrying an hourglass stretched upward towards the kneeling figure of the pope.
Archpriests since 1053
List of archpriests of the Vatican Basilica:
Mauro Gambetti (20 February 2021 – present)
Specifications
See also
Architecture of cathedrals and great churches
Architecture of Rome
History of early modern period domes
List of tallest domes
Index of Vatican City-related articles
List of basilicas in Italy
List of oldest church buildings
List of Roman Catholic basilicas
List of tallest buildings in Rome
Notes
References
Bibliography
External links
Vaticanstate.va – official website with images and information about the Basilica
"Virtual Reality Tour of the Basilica of Saint Peter"
St Peter's Basilica.info ‒ unofficial website on the basilica, with images and text from different books.
360 Degree Photographs Inside Saint Peter's Basilica
Google Maps: Vatican ‒ Satellite image of the Basilica
Peter
Coronation church buildings
Church buildings with domes
Terminating vistas
Renaissance architecture in Rome
Baroque architecture in Rome
Tourist attractions in Vatican City
Churches completed in 1626
Tombs of apostles
Burial places of popes
1626 establishments in the Papal States
Pope Julius II
Donato Bramante church buildings
Michelangelo church buildings
Gian Lorenzo Bernini church buildings
St Peter's
17th-century Roman Catholic church buildings in Italy |
null | null | Mangrove | eng_Latn | A mangrove is a shrub or tree that grows in coastal saline or brackish water. The term is also used for tropical coastal vegetation consisting of such species. Mangroves occur worldwide in the tropics and subtropics and even some temperate coastal areas, mainly between latitudes 30° N and 30° S, with the greatest mangrove area within 5° of the equator.
Mangrove plant families first appeared during the Late Cretaceous to Paleocene epochs, and became widely distributed in part due to the movement of tectonic plates. The oldest known fossils of mangrove palm date to 75 million years ago.
Mangroves are salt-tolerant trees, also called halophytes, and are adapted to live in harsh coastal conditions. They contain a complex salt filtration system and
a complex root system to cope with saltwater immersion and wave action. They are adapted to the low-oxygen conditions of waterlogged mud, but are most likely to thrive in the upper half of the intertidal zone.
The mangrove biome, often called the mangrove forest or mangal, is a distinct saline woodland or shrubland habitat characterized by depositional coastal environments, where fine sediments (often with high organic content) collect in areas protected from high-energy wave action. The saline conditions tolerated by various mangrove species range from brackish water, through pure seawater (3 to 4% salinity), to water concentrated by evaporation to over twice the salinity of ocean seawater (up to 9% salinity).
Beginning in 2010
remote sensing technologies and global data have been used to assess areas, conditions and deforestation rates of mangroves around the world. In 2018, the Global Mangrove Watch Initiative released a new global baseline which estimates the total mangrove forest area of the world as of 2010 at , spanning 118 countries and territories. Mangrove loss continues due to human activity, with a global annual deforestation rate estimated at 0.16%, and per-country rates as high as 0.70%. Degradation in quality of remaining mangroves is also an important concern.
There is interest in mangrove restoration for several reasons. Mangroves support sustainable coastal and marine ecosystems. They protect nearby areas from tsunamis and extreme weather events. Mangrove forests are also effective at carbon sequestration and storage and mitigate climate change. As the effects of climate change become more severe, mangrove ecosystems are expected to help local ecosystems adapt and be more resilient to changes like extreme weather and sea level rise. The success of mangrove restoration may depend heavily on engagement with local stakeholders, and on careful assessment to ensure that growing conditions will be suitable for the species chosen.
Etymology
Etymology of the English term mangrove can only be speculative and is disputed.
The term may have come to English from the Portuguese or the Spanish
. Farther back, it may be traced to South America and Cariban and Arawakan languages such as Taíno. Other possibilities include the Malay language and the Guarani language.
The English usage may reflect a corruption via folk etymology of the words mangrow and grove.
The word "mangrove" is used in at least three senses:
most broadly to refer to the habitat and entire plant assemblage or mangal, for which the terms mangrove forest biome and mangrove swamp are also used;
to refer to all trees and large shrubs in a mangrove swamp; and
narrowly to refer only to mangrove trees of the genus Rhizophora of the family Rhizophoraceae.
Biology
Of the recognized 110 mangrove species, only about 54 species in 20 genera from 16 families constitute the "true mangroves", species that occur almost exclusively in mangrove habitats. Demonstrating convergent evolution, many of these species found similar solutions to the tropical conditions of variable salinity, tidal range (inundation), anaerobic soils, and intense sunlight. Plant biodiversity is generally low in a given mangrove. The greatest biodiversity of mangroves occurs in Southeast Asia, particularly in the Indonesian archipelago.
Adaptations to low oxygen
The red mangrove (Rhizophora mangle) survives in the most inundated areas, props itself above the water level with stilt or prop roots and then absorbs air through lenticels in its bark.
The black mangrove (Avicennia germinans) lives on higher ground and develops many specialized root-like structures called pneumatophores, which stick up out of the soil like straws for breathing.
These "breathing tubes" typically reach heights of up to , and in some species, over . The four types of pneumatophores are stilt or prop type, snorkel or peg type, knee type, and ribbon or plank type. Knee and ribbon types may be combined with buttress roots at the base of the tree. The roots also contain wide aerenchyma to facilitate transport within the plants.
Nutrient uptake
Because the soil is perpetually waterlogged, little free oxygen is available. Anaerobic bacteria liberate nitrogen gas, soluble ferrum (iron), inorganic phosphates, sulfides, and methane, which make the soil much less nutritious. Pneumatophores (aerial roots) allow mangroves to absorb gases directly from the atmosphere, and other nutrients such as iron, from the inhospitable soil. Mangroves store gases directly inside the roots, processing them even when the roots are submerged during high tide.
Limiting salt intake
Red mangroves exclude salt by having significantly impermeable roots which are highly suberised (impregnated with suberin), acting as an ultra-filtration mechanism to exclude sodium salts from the rest of the plant. Analysis of water inside mangroves has shown 90% to 97% of salt has been excluded at the roots. In a frequently cited concept that has become known as the "sacrificial leaf", salt which does accumulate in the shoot (sprout) then concentrates in old leaves, which the plant then sheds. However, recent research suggests the older, yellowing leaves have no more measurable salt content than the other, greener leaves. Red mangroves can also store salt in cell vacuoles. White and grey mangroves can secrete salts directly; they have two salt glands at each leaf base (correlating with their name—they are covered in white salt crystals).
Limiting water loss
Because of the limited fresh water available in salty intertidal soils, mangroves limit the amount of water they lose through their leaves. They can restrict the opening of their stomata (pores on the leaf surfaces, which exchange carbon dioxide gas and water vapor during photosynthesis). They also vary the orientation of their leaves to avoid the harsh midday sun and so reduce evaporation from the leaves. A captive red mangrove grows only if its leaves are misted with fresh water several times a week, simulating frequent tropical rainstorms.<ref>{{cite web |author=Calfo, A. |year=2006 |url=Mangroves for the Marine Aquarium by Anthony Calfo - Reefkeeping.com |title=Mangroves for the Marine Aquarium |publisher=Reefkeeping |access-date=2012-02-08}}</ref>
Filtration of seawater
A 2016 study by Kim et al. investigated the biophysical characteristics of sea water filtration in the roots of the mangrove Rhizophora stylosa from a plant hydrodynamic point of view. R. stylosa can grow even in saline water and the salt level in its roots is regulated within a certain threshold value through filtration. The root possesses a hierarchical, triple layered pore structure in the epidermis and most Na+ ions are filtered at the first sublayer of the outermost layer. The high blockage of Na+ ions is attributed to the high surface zeta potential of the first layer. The second layer, which is composed of macroporous structures, also facilitates Na+ ion filtration. The study provides insights into the mechanism underlying water filtration through halophyte roots and could serve as a basis for the development of a bio-inspired method of desalination.
Uptake of Na+ ions is desirable for halophytes to build up osmotic potential, absorb water and sustain turgor pressure. However, excess Na+ions may work on toxic element. Therefore, halophytes try to adjust salinity delicately between growth and survival strategies. In this point of view, a novel sustainable desalination method can be derived from halophytes, which are in contact with saline water through their roots. Halophytes exclude salt through their roots, secrete the accumulated salt through their aerial parts and sequester salt in senescent leaves and/or the bark. Mangroves are facultative halophytes and Bruguiera is known for its special ultrafiltration system that can filter approximately 90% of Na+ions from the surrounding seawater through the roots. The species also exhibits a high rate of salt rejection. The water-filtering process in mangrove roots has received considerable attention for several decades. Morphological structures of plants and their functions have been evolved through a long history to survive against harsh environmental conditions.
Increasing survival of offspring
In this harsh environment, mangroves have evolved a special mechanism to help their offspring survive. Mangrove seeds are buoyant and are therefore suited to water dispersal. Unlike most plants, whose seeds germinate in soil, many mangroves (e.g. red mangrove) are viviparous, meaning their seeds germinate while still attached to the parent tree. Once germinated, the seedling grows either within the fruit (e.g. Aegialitis, Avicennia and Aegiceras), or out through the fruit (e.g. Rhizophora, Ceriops, Bruguiera and Nypa) to form a propagule (a ready-to-go seedling) which can produce its own food via photosynthesis.
The mature propagule then drops into the water, which can transport it great distances. Propagules can survive desiccation and remain dormant for over a year before arriving in a suitable environment. Once a propagule is ready to root, its density changes so that the elongated shape now floats vertically rather than horizontally. In this position, it is more likely to lodge in the mud and root. If it does not root, it can alter its density and drift again in search of more favorable conditions.
Taxonomy and evolution
The following listings, based on Tomlinson, 2016, give the mangrove species in each listed plant genus and family. Mangrove environments in the Eastern Hemisphere harbor six times as many species of trees and shrubs as do mangroves in the New World. Genetic divergence of mangrove lineages from terrestrial relatives, in combination with fossil evidence, suggests mangrove diversity is limited by evolutionary transition into the stressful marine environment, and the number of mangrove lineages has increased steadily over the Tertiary with little global extinction.
True mangroves
Minor components
Species distribution
Mangroves are a type of tropical vegetation with some outliers established in subtropical latitudes, notably in South Florida and southern Japan, as well as South Africa, New Zealand and Victoria (Australia). These outliers result either from unbroken coastlines and island chains or from reliable supplies of propagules floating on warm ocean currents from rich mangrove regions.
"At the limits of distribution, the formation is represented by scrubby, usually monotypic Avicennia-dominated vegetation, as at Westonport Bay and Corner Inlet, Victoria, Australia. The latter locality is the highest latitude (38° 45'S) at which mangroves occur naturally. The mangroves in New Zealand, which extend as far south as 37°, are of the same type; they start as low forest in the northern part of the North Island but become low scrub toward their southern limit. In both instances, the species is referred to as Avicennia marina var. australis, although genetic comparison is clearly needed. In Western Australia, A. marina extends as far south as Sunbury (33° 19'S). In the northern hemisphere, scrubby Avicennia gerrninans in Florida occurs as far north as St. Augustine on the east coast and Cedar Point on the west. There are records of A. germinans and Rhizophora mangle for Bermuda, presumably supplied by the Gulf Stream. In southern Japan, Kandelia obovata occurs to about 31 °N (Tagawa in Hosakawa et al., 1977, but initially referred to as K. candel)."
Mangrove forests
Mangrove forests, also called mangrove swamps or mangals, are found in tropical and subtropical tidal areas. Areas where mangroves occur include estuaries and marine shorelines.
The intertidal existence to which these trees are adapted represents the major limitation to the number of species able to thrive in their habitat. High tide brings in salt water, and when the tide recedes, solar evaporation of the seawater in the soil leads to further increases in salinity. The return of tide can flush out these soils, bringing them back to salinity levels comparable to that of seawater.
At low tide, organisms are also exposed to increases in temperature and reduced moisture before being then cooled and flooded by the tide. Thus, for a plant to survive in this environment, it must tolerate broad ranges of salinity, temperature, and moisture, as well as several other key environmental factors—thus only a select few species make up the mangrove tree community.
About 110 species are considered mangroves, in the sense of being trees that grow in such a saline swamp, though only a few are from the mangrove plant genus, Rhizophora. However, a given mangrove swamp typically features only a small number of tree species. It is not uncommon for a mangrove forest in the Caribbean to feature only three or four tree species. For comparison, the tropical rainforest biome contains thousands of tree species, but this is not to say mangrove forests lack diversity. Though the trees themselves are few in species, the ecosystem that these trees create provides a home (habitat) for a great variety of other species, including as many as 174 species of marine megafauna.
Mangrove plants require a number of physiological adaptations to overcome the problems of low environmental oxygen levels, high salinity, and frequent tidal flooding. Each species has its own solutions to these problems; this may be the primary reason why, on some shorelines, mangrove tree species show distinct zonation. Small environmental variations within a mangal may lead to greatly differing methods for coping with the environment. Therefore, the mix of species is partly determined by the tolerances of individual species to physical conditions, such as tidal flooding and salinity, but may also be influenced by other factors, such as crabs preying on plant seedlings.
Once established, mangrove roots provide an oyster habitat and slow water flow, thereby enhancing sediment deposition in areas where it is already occurring. The fine, anoxic sediments under mangroves act as sinks for a variety of heavy (trace) metals which colloidal particles in the sediments have concentrated from the water. Mangrove removal disturbs these underlying sediments, often creating problems of trace metal contamination of seawater and organisms of the area.
Mangrove swamps protect coastal areas from erosion, storm surge (especially during tropical cyclones), and tsunamis. They limit high-energy wave erosion mainly during events such as storm surges and tsunamis.
The mangroves' massive root systems are efficient at dissipating wave energy. Likewise, they slow down tidal water enough so that its sediment is deposited as the tide comes in, leaving all except fine particles when the tide ebbs. In this way, mangroves build their environments. Because of the uniqueness of mangrove ecosystems and the protection against erosion they provide, they are often the object of conservation programs, including national biodiversity action plans.
The unique ecosystem found in the intricate mesh of mangrove roots offers a quiet marine habitat for young organisms. In areas where roots are permanently submerged, the organisms they host include algae, barnacles, oysters, sponges, and bryozoans, which all require a hard surface for anchoring while they filter-feed. Shrimps and mud lobsters use the muddy bottoms as their home. Mangrove crabs eat the mangrove leaves, adding nutrients to the mangal mud for other bottom feeders. In at least some cases, the export of carbon fixed in mangroves is important in coastal food webs.
Mangrove plantations in Vietnam, Thailand, Philippines, and India host several commercially important species of fish and crustaceans.
Mangrove forests can decay into peat deposits because of fungal and bacterial processes as well as by the action of termites. It becomes peat in good geochemical, sedimentary, and tectonic conditions. The nature of these deposits depends on the environment and the types of mangroves involved. In Puerto Rico, the red, white, and black mangroves occupy different ecological niches and have slightly different chemical compositions, so the carbon content varies between the species, as well between the different tissues of the plant (e.g., leaf matter versus roots).
In Puerto Rico, there is a clear succession of these three trees from the lower elevations, which are dominated by red mangroves, to farther inland with a higher concentration of white mangroves. Mangrove forests are an important part of the cycling and storage of carbon in tropical coastal ecosystems. Knowing this, scientists seek to reconstruct the environment and investigate changes to the coastal ecosystem over thousands of years using sediment cores. However, an additional complication is the imported marine organic matter that also gets deposited in the sediment due to the tidal flushing of mangrove forests.
Termites play an important role in the formation of peat from mangrove materials. They process fallen leaf litter, root systems and wood from mangroves into peat to build their nests. Termites stabilise the chemistry of this peat and represent approximately 2% of above ground carbon storage in mangroves. As the nests are buried over time this carbon is stored in the sediment and the carbon cycle continues.
Mangroves are an important source of blue carbon. Globally, mangroves stored of carbon in 2012. Two percent of global mangrove carbon was lost between 2000 and 2012, equivalent to a maximum potential of of CO2 emissions.
Globally, mangroves have been shown to provide measurable economic protections to coastal communities affected by tropical storms.
Mangrove microbiomes
Plant microbiomes play crucial roles in their health and productivity of mangroves. Many researchers have successfully applied knowledge acquired about plant microbiomes to produce specific inocula for crop protection. Such inocula can stimulate plant growth by releasing phytohormones and enhancing uptake of some mineral nutrients (particularly phosphorus and nitrogen). However, most of the plant microbiome studies have focused on the model plant Arabidopsis thaliana and economically important crop plants, such as rice , barley, wheat, maize and soybean. There is less information on microbiomes of tree species. Plant microbiomes are determined by plant-related factors (e.g., genotype, organ, species, and health status) and environmental factors (e.g., land use, climate, and nutrient availability). Two of the plant-related factors, plant species and genotypes, have been shown to play significant roles in shaping rhizosphere and plant microbiomes, as tree genotypes and species are associated with specific microbial communities. Different plant organs also have specific microbial communities depending on plant-associated factors (plant genotype, available nutrients, and organ-specific physicochemical conditions) and/or environmental conditions (associated with aboveground and underground surfaces and disturbances).
Root microbiome
Mangrove roots harbour a repertoire of microbial taxa that contribute to important ecological functions in mangrove ecosystems. Similar to typical terrestrial plants, mangroves depend upon mutually beneficial interactions with microbial communities. In particular, microbes residing in developed roots could help mangroves transform nutrients into usable forms prior to plant assimilation. These microbes also provide mangroves phytohormones for suppressing phytopathogens or helping mangroves withstand heat and salinity. In turn, root-associated microbes receive carbon metabolites from the plant via root exudates, thus close associations between the plant and microbes are established for their mutual benefits. Material was copied from this source, which is available under a Creative Commons Attribution 4.0 International License.
Highly diverse microbial communities (mainly bacteria and fungi) have been found to inhabit and function in mangrove roots. For example, diazotrophic bacteria in the vicinity of mangrove roots could perform biological nitrogen fixation, which provides 40–60% of the total nitrogen required by mangroves; the soil attached to mangrove roots lacks oxygen but is rich in organic matter, providing an optimal microenvironment for sulfate-reducing bacteria and methanogens, ligninolytic, cellulolytic, and amylolytic fungi are prevalent in the mangrove root environment; rhizosphere fungi could help mangroves survive in waterlogged and nutrient-restricted environments. These studies have provided increasing evidences to support the importance of root-associated bacteria and fungi for mangrove growth and health.
Recent studies have investigated the detailed structure of root-associated microbial communities at a continuous fine-scale in other plants, where a microhabitat was divided into four root compartments: endosphere, episphere, rhizosphere, and nonrhizosphere. Moreover, the microbial communities in each compartment have been reported to have unique characteristics. The rhizosphere could emit root exudates that selectively enriched specific microbial populations; however, these exudates were found to exert only marginal impacts on microbes in the nonrhizosphere soil. Furthermore, it was noted that the root episphere, rather than the rhizosphere, was primarily responsible for controlling the entry of specific microbial populations into the root, resulting in the selective enrichment of Proteobacteria in the endosphere. These findings provide new insights into the niche differentiation of root-associated microbial communities, Nevertheless, amplicon-based community profiling may not provide the functional characteristics of root-associated microbial communities in plant growth and biogeochemical cycling. Unraveling functional patterns across the four root compartments holds a great potential for understanding functional mechanisms responsible for mediating root–microbe interactions in support of enhancing mangrove ecosystem functioning.
Mangrove viromes
Mangroves forests are one of the most carbon-rich biomes, accounting for 11% of the total input of terrestrial carbon into oceans. Although viruses are thought to significantly influence local and global biogeochemical cycles, though as of 2019 little information was available about the community structure, genetic diversity and ecological roles of viruses in mangrove ecosystems.
Viruses are the most abundant biological entities on earth, present in virtually all ecosystems. By lysing their hosts, that is, by rupturing their cell membranes, viruses control host abundance and affect the structure of host communities. Viruses also influence their host diversity and evolution through horizontal gene transfer, selection for resistance and manipulation of bacterial metabolisms. Importantly, marine viruses affect local and global biogeochemical cycles through the release of substantial amounts of organic carbon and nutrients from hosts and assist microbes in driving biogeochemical cycles with auxiliary metabolic genes (AMGs).
It is presumed AMGs augment viral-infected host metabolism and facilitate the production of new viruses. AMGs have been extensively explored in marine cyanophages and include genes involved in photosynthesis, carbon turnover, phosphate uptake and stress response. Cultivation-independent metagenomic analysis of viral communities has identified additional AMGs that are involved in motility, central carbon metabolism, photosystem I, energy metabolism, iron–sulphur clusters, anti-oxidation and sulphur and nitrogen cycling. Interestingly, a recent analysis of Pacific Ocean Virome data identified niche-specialised AMGs that contribute to depth-stratified host adaptations. Given that microbes drive global biogeochemical cycles, and a large fraction of microbes is infected by viruses at any given time, viral-encoded AMGs must play important roles in global biogeochemistry and microbial metabolic evolution.
Mangrove forests are the only woody halophytes that live in salt water along the world’s subtropical and tropical coastlines. Mangroves are one of the most productive and ecologically important ecosystems on earth. The rates of primary production of mangroves equal those of tropical humid evergreen forests and coral reefs. As a globally relevant component of the carbon cycle, mangroves sequester approximately 24 million metric tons of carbon each year. Most mangrove carbon is stored in soil and sizable belowground pools of dead roots, aiding in the conservation and recycling of nutrients beneath forests. Although mangroves cover only 0.5% of the earth’s coastal area, they account for 10–15% of the coastal sediment carbon storage and 10–11% of the total input of terrestrial carbon into oceans. The disproportionate contribution of mangroves to carbon sequestration is now perceived as an important means to counterbalance greenhouse gas emissions.
Despite the ecological importance of mangrove ecosystem, our knowledge on mangrove biodiversity is notably limited. Previous reports mainly investigated the biodiversity of mangrove fauna, flora and bacterial communities. Particularly, little information is available about viral communities and their roles in mangrove soil ecosystems. In view of the importance of viruses in structuring and regulating host communities and mediating element biogeochemical cycles, exploring viral communities in mangrove ecosystems is essential. Additionally, the intermittent flooding of sea water and resulting sharp transition of mangrove environments may result in substantially different genetic and functional diversity of bacterial and viral communities in mangrove soils compared with those of other systems.
Genome sequencing
Rhizophoreae as revealed by whole-genome sequencing
See also
Coastal management
Mangrove swamp
Mangrove restoration
Salt marsh
Longshore drift
Coastal erosion
Coastal geography
Ecological values of mangrove
Blue carbon
Keystone species
References
Further reading
Saenger, Peter (2002). Mangrove Ecology, Silviculture, and Conservation. Kluwer Academic Publishers, Dordrecht. .
Thanikaimoni, Ganapathi (1986). Mangrove Palynology UNDP/UNESCO and the French Institute of Pondicherry, ISSN 0073-8336 (E).
Tomlinson, Philip B. (1986). The Botany of Mangroves. Cambridge University Press, Cambridge, .
Teas, H. J. (1983). Biology and Ecology of Mangroves. W. Junk Publishers, The Hague. .
Agrawala, Shardul; Hagestad; Marca; Koshy, Kayathu; Ota, Tomoko; Prasad, Biman; Risbey, James; Smith, Joel; Van Aalst, Maarten. 2003. Development and Climate Change in Fiji: Focus on Coastal Mangroves. Organisation of Economic Co-operation and Development, Paris, Cedex 16, France.
Glenn, C. R. 2006. "Earth's Endangered Creatures"
Twilley, R. R., V.H. Rivera-Monroy, E. Medina, A. Nyman, J. Foret, T. Mallach, and L. Botero. 2000. Patterns of forest development in mangroves along the San Juan River estuary, Venezuela. Forest Ecology and Management
Spalding, Mark; Kainuma, Mami and Collins, Lorna (2010) World Atlas of Mangroves'' Earthscan, London, ; 60 maps showing worldwide mangrove distribution
External links
Top 10 Mangrove Forest In The World - Travel Mate
*
In May 2011, the VOA Special English service of the Voice of America broadcast a 15-minute program on mangrove forests. A transcript and MP3 of the program, intended for English learners, can be found at Mangrove Forests Could Be a Big Player in Carbon Trading
Queensland’s coastal kidneys: mangroves. Stacey Larner, John Oxley Library Blog. State Library of Queensland.
Aquatic biomes
Aquatic ecology
Blue carbon
Terrestrial biomes
Plant common names
Oceanographical terminology |
null | null | Grand Moff Tarkin | eng_Latn | Grand Moff Wilhuff Tarkin is a fictional character in the Star Wars franchise, introduced in the original 1977 Star Wars film (played by Peter Cushing) as one of its two central antagonists, alongside Darth Vader. In the film, Tarkin is depicted as a high-ranking officer of the Galactic Empire, placed in charge of the operations on the Death Star, the Empire's dwarf planet-sized battle station. He orders the destruction of the planet Alderaan by the station's superlaser, and is killed at the end of the film, when Luke Skywalker destroys the Death Star.
Outside of the original film, Tarkin has made appearances in various canon and non-canon Star Wars media, including the animated series The Clone Wars, Rebels, and The Bad Batch, and the films Revenge of the Sith and Rogue One. He is also the protagonist of the 2014 novel Tarkin, which details the character's backstory and rise to power within the Empire's ranks. In the novel, Tarkin advocates for the Tarkin Doctrine, an entirely fear-based military system in which the Empire uses large numbers of intimidating capital ships and strike fighters to seem undefeatable, so that no one would dare to try and rise up against it. His philosophy of "ruling through fear of force" becomes central to Imperial policy and earns him the position of the first Grand Moff in the Empire.
Tarkin's character has been well-received, and he has been called "one of the most formidable villains in Star Wars history."
Concept and creation
Grand Moff Tarkin's character was originally conceived as a holy man from the planet Aquila, but was changed later in the creative process into an antagonist. As the Emperor would not appear until later in the original trilogy, Lucas used Tarkin's final version as the "main villain" of the first film, a personification of the Empire. According to a book created to help promote the original film to prospective theaters, he aspires to become the Emperor.
Portrayals
During the production of Star Wars, Peter Cushing found Tarkin's boots, furnished by the wardrobe department, to be very uncomfortable. Director George Lucas agreed to limit shots where Cushing's feet would be visible, allowing him to wear slippers. Cushing admitted many years later to not knowing what a "Grand Moff" actually was, joking that it sounded like "something which infests a clothes closet" and decided to play him as a "deeply cross and unpleasant gentleman".
In the 2005 prequel film Revenge of the Sith, Wayne Pygram was able to achieve the likeness of a young version of Tarkin through the use of prosthetic makeup. For his performance as Tarkin in the animated series The Clone Wars, voice actor Stephen Stanton researched Cushing's performances and then tried to imitate what Cushing might have sounded like in his mid-thirties and soften his voice to portray a level of humanity.
In the 2016 anthology film Rogue One, archive footage and a digital scan of Peter Cushing's life mask made for the 1984 film Top Secret! were used to create a 3D CGI mask which was augmented and mapped to actor Guy Henry's face. Henry had studied Cushing's mannerisms many years previously for the lead role in British TV show Young Sherlock, but insisted on a screen test as he was not comfortable that his vocal imitation was accurate, stating he sounded more "Peter O'Toole than Peter Cushing". The ILM team searched through hours of footage to find suitable material of Cushing to build from. The footage from A New Hope was lit very differently to the lighting used in Rogue One and had to be digitally changed. The more they manipulated the lighting to match the other actors in the scenes the less like Cushing the character model looked, which meant creating a balancing act between "a digital figure" and "one who looked precisely like Cushing". The owners of Cushing's estate were heavily involved with the creation and had input right down to "small, subtle adjustments". The result, which has been called "one of the most complex and costly CGI re-creations ever", received a mixed response, with questions being raised about the morality of using a dead actor's likeness.
Appearances
Film
Star Wars (1977)
Introduced in the first film in the original Star Wars trilogy, Grand Moff Tarkin is the Governor of the Imperial Outland Regions and commander of the Death Star. After Emperor Palpatine dissolves the Imperial Senate, Tarkin and Darth Vader (portrayed by David Prowse, voiced by James Earl Jones) are charged with pursuing and destroying the Rebel Alliance. He threatens Princess Leia Organa (Carrie Fisher) with the destruction of her home planet, Alderaan, if Leia does not reveal the location of the Rebel main base of operations. When Leia names the planet Dantooine as the base's location, he destroys Alderaan regardless, hoping to make an example out of the planet's support of the Rebellion. Upon learning that Leia's coerced information was false, Tarkin orders Leia's execution.
He allows the Rebels to escape the Death Star with Leia after placing a tracking beacon on the Millennium Falcon in order to find the Rebel base. He orders the Death Star to destroy the Rebel base on Yavin 4. In the film's climax, Tarkin refuses to believe that the Death Star is in danger from the Rebel starfighter attack; as a result, he refuses to evacuate due to his arrogance. He is subsequently killed by Luke Skywalker (Mark Hamill), who succeeds in destroying the Death Star by firing torpedoes into the exhaust port.
Revenge of the Sith (2005)
At the end of Star Wars: Episode III – Revenge of the Sith, the final film in the Star Wars prequel trilogy, a younger version of Tarkin, played by Wayne Pygram, makes a cameo appearance overseeing the original Death Star's construction, standing beside Darth Vader (Hayden Christensen) and Emperor Palpatine (Ian McDiarmid)
Rogue One (2016)
In the film Rogue One (which takes place just before the events in A New Hope), Orson Krennic (Ben Mendelsohn), Director of Advanced Weapons Research for the Imperial Military, meets with Tarkin, who expresses his skepticism about Krennic's management of the Death Star project. Tarkin oversees the Death Star's first attack on the Rebellion when it is used to destroy the Holy City of the planet Jedha. Impressed, he congratulates Krennic before announcing that he is going to take command of the Death Star from that point on, pointing out security breaches that had occurred under Krennic's command (much to Krennic's chagrin). Later in the film, Tarkin is informed of a Rebel attack on Scarif, the planet where the plans to the Death Star are kept. He orders the jump to hyperspace to the planet, where an ongoing battle between the Empire and the Rebel Alliance rages. Tarkin has the Death Star target and destroy the Scarif base, killing Krennic, Jyn Erso (Felicity Jones), Cassian Andor (Diego Luna), and any other survivor of the ground battle.
Television series
The Clone Wars (2010–2013)
In the animated television series Star Wars: The Clone Wars, set during the events of the prequel trilogy, a younger Tarkin (voiced by Stephen Stanton) is depicted as a Captain and later an Admiral in the Galactic Republic Navy.
In the third season, Captain Tarkin and Jedi Master Even Piell (Blair Bess) are ambushed and attacked by Separatist forces. Prisoners to the Citadel, Tarkin and Piell are freed from captivity by a rescue team. Initially pessimistic about being in enemy territory, Tarkin puts himself at odds with Jedi Knight Anakin Skywalker (Matt Lanter), but their respective opinions of each other improve when each realizes that both know Supreme Chancellor Palpatine (Ian Abercrombie). During a skirmish, Tarkin fights and attempts to execute Citadel Warden Osi Sobeck (James Arnold Taylor), but fails when Sobeck swiftly retaliates and nearly kills him. However, Tarkin is rescued just in time by Anakin's Padawan Ahsoka Tano (Ashley Eckstein).
In the fifth season, Tarkin, now an admiral, suspects Ahsoka of masterminding a terrorist attack on the Jedi Temple, and attempts to have the Padawan arrested. After Ahsoka is recaptured, Tarkin requests that the Jedi Order expel her and turn her over to the Republic so she can receive more "impartial" judgement. The Jedi concede and Ahsoka is tried before a jury of senators, with Tarkin heading the prosecution while Padmé Amidala (Catherine Taber) heads the defense. Despite Padmé's impressive defense, Tarkin casts doubt by mentioning that Ahsoka had been seen with Dark Jedi Asajj Ventress (Nika Futterman). After Tarkin and Padmé's arguments conclude and the jury reaches a verdict that Palpatine is about to read, Anakin arrives with Barriss Offee (Olivia d'Abo), the real mastermind of the attack.
Star Wars Rebels (2014–2018)
In the television series Star Wars Rebels, Tarkin (voiced once again by Stanton) starts off as Governor of the Outer Rim territories, including Lothal, but eventually receives his Grand Moff title. He visits the planet Lothal to deal with its growing Rebel activity, and reprimands Minister Maketh Tua (Kath Soucie), Agent Kallus (David Oyelowo), and the Inquisitor for their repeated failures to stop the planet's Rebel cell. Tarkin has the Inquisitor execute Commandant Aresko and Taskmaster Grint (both David Shaughnessy) for their inability to deal with the cell's leader, Jedi Knight Kanan Jarrus (Freddie Prinze Jr.). Later, Tarkin sets a trap for the Rebels and manages to capture Kanan during their mission to send a message through one of the planet's communication towers. The Rebels' message gets sent out, but Tarkin then orders the tower to be destroyed. Kanan is tortured by the Inquisitor and transported to the Mustafar system aboard Tarkin's Star Destroyer. During Kanan's rescue by Rebel forces, Tarkin's Star Destroyer is destroyed, and the Inquisitor is killed. On Lothal, Tarkin introduces Agent Kallus to Darth Vader (James Earl Jones).
Taking the loss of his Star Destroyer personally, Tarkin orders that Maketh Tua be brought before him for her failure to find the Lothal rebels. Knowing the true meaning behind the summoning after the deaths of Aresko and Grint, Tua attempts to defect, but she is killed before she can leave Lothal. Tarkin later appears in the season three premiere where he meets with Governor Pryce (Mary Elizabeth Glynn) regarding the Lothal Rebels. He subsequently grants her request for use of the Seventh Fleet, commanded by Grand Admiral Thrawn (Lars Mikkelsen). Tarkin appears via hologram in the season three finale, being briefed on Thrawn's imminent attack on Atollon to stymie a coordinated Rebel attack on Lothal. Towards the end of season four, he warns Thrawn that he must prove the worth of his TIE Defender program, Thrawn's prototype design of a new Imperial Star Fighter, or it will be shut down in favor of diverting resources to Krennic's "Stardust project" - which would eventually become the Death Star.
Star Wars: The Bad Batch (2021)
Tarkin (voiced again by Stanton) appears in the Disney+ series Star Wars: The Bad Batch, set between the events of The Clone Wars and Rebels. In the premiere episode, "Aftermath", Admiral Tarkin arrives on Kamino to discuss with Prime Minister Lama Su if the Empire should continue to produce clone troopers, noting that conscripted soldiers would be less expensive. He later evaluates Clone Force 99, a group of genetically enhanced clone troopers also known as the Bad Batch, through a battle simulation. While impressed by their extraordinary skills, he also notices disobedience in them. He then dispatches them to the planet of Onderon to eliminate a group of refugees led by Saw Gerrera, as a test of their loyalty. When the Bad Batch disobeys orders, Tarkin has them arrested, but separates Bad Batch member Crosshair from the others after noticing that he is more loyal to the Empire. Tarkin has Kaminoan scientist Nala Se intensify the programming of Crosshair's inhibitor chip, turning him against his teammates, who escape Kamino.
Tarkin is later promoted to Governor, and in the episode "Replacements", he puts Crosshair in charge of an "elite squad" of recruited soldiers, who are then sent to finish the mission the Bad Batch had abandoned. After the mission is successful, Tarkin sees potential for conscripts. In the episode "Return to Kamino", after Kamino itself is evacuated, Tarkin orders Vice Admiral Rampart to destroy the cloning facility.
Novels
Tarkin appears in Catalyst: A Rogue One Novel, the prequel novel to Rogue One, where Tarkin forms a rivalry with Advanced Weapons Director Orson Krennic.
Tarkin (2014)
Star Wars: Tarkin explores the title character's origins, and chronicles how he meets and aligns himself with Emperor Palpatine and Darth Vader prior to the events of A New Hope. The novel was one of the first four canon novels to be released in 2014 and 2015. Tarkin's Star Destroyer, the Executrix, is introduced; it later appears in Rogue One.
From a Certain Point of View (2017)
The From a Certain Point of View anthology short story "Of MSE-6 and Men", written by Glen Weldon, Tarkin is revealed to have been sleeping with the stormtrooper TK-421 during the events of A New Hope, after capturing Leia Organa, but before his armour was stolen by Luke Skywalker in order to rescue her (Leia).
Legends
With the 2012 acquisition of Lucasfilm by The Walt Disney Company, most of the licensed Star Wars novels and comics produced since the originating 1977 film Star Wars were rebranded as Star Wars Legends and declared non-canon to the franchise in April 2014.
In the Star Wars Expanded Universe, Tarkin appears in several novels, including Death Star, Cloak of Deception, Rogue Planet and Dark Lord: The Rise of Darth Vader. The original Marvel Star Wars comic series features a powerful superweapon called "The Tarkin" after the late Grand Moff. His protegee and lover, Natasi Daala, is later featured as a frequent antagonist in novels set after his death.
West End Games' roleplaying material describes the "Tarkin Doctrine", which emphasizes ruling "through the fear of force, rather than force itself", and has been mentioned various other times in the Star Wars canon. He is also mentioned during the New Jedi Order novel series as having been present on Zonama Sekot with Anakin Skywalker. In the comics series Darth Vader and the Lost Command, Tarkin is seen telling Vader of a missing convoy of Imperial craft and adds that his own son was in command of the convoy and is also missing.
References
Footnotes
Citations
External links
Grand Moff Tarkin on IMDb
Characters created by George Lucas
Star Wars CGI characters
Fictional admirals
Fictional bisexual males
Fictional governors
Fictional henchmen
Fictional mass murderers
Fictional military strategists
Fictional war veterans
Film characters introduced in 1977
Male film villains
Star Wars Anthology characters
Star Wars Skywalker Saga characters
Star Wars: The Clone Wars characters
Star Wars Rebels characters
Star Wars literary characters
Film supervillains |
null | null | Anaphylaxis | eng_Latn | Anaphylaxis, often known as anaphylactic shock, is a serious allergic reaction that is rapid in onset and may cause death. It typically causes more than one of the following: an itchy rash, throat or tongue swelling, shortness of breath, vomiting, lightheadedness and low blood pressure. These symptoms typically come on over minutes to hours.
Common causes include insect bites and stings, foods, and medications. Other causes include latex exposure and exercise; cases may also occur without an obvious reason. The mechanism involves the release of inflammatory mediators from certain types of white blood cells triggered by either immunologic or non-immunologic mechanisms. Diagnosis is based on the presenting symptoms and signs after exposure to a potential allergen.
The primary treatment of anaphylaxis is epinephrine injection into a muscle, intravenous fluids, then placing the person "in a reclining position with feet elevated to help restore normal blood flow". Additional doses of epinephrine may be required. Other measures, such as antihistamines and steroids, are complementary. Carrying an epinephrine autoinjector and identification regarding the condition is recommended in people with a history of anaphylaxis.
Worldwide, 0.05–2% of the population is estimated to experience anaphylaxis at some point in life. Globally, as underreporting declined into the 2010s, the rate appeared to be increasing. It occurs most often in young people and females. About 99.7% of people hospitalized with anaphylaxis in the United States survive. The term comes from the , and the .
Signs and symptoms
Anaphylaxis typically presents many different symptoms over minutes or hours with an average onset of 5 to 30 minutes if exposure is intravenous and up to 2 hours if from eating food. The most common areas affected include: skin (80–90%), respiratory (70%), gastrointestinal (30–45%), heart and vasculature (10–45%), and central nervous system (10–15%) with usually two or more being involved.
Skin
Symptoms typically include generalized hives, itchiness, flushing, or swelling (angioedema) of the afflicted tissues. Those with angioedema may describe a burning sensation of the skin rather than itchiness. Swelling of the tongue or throat occurs in up to about 20% of cases. Other features may include a runny nose and swelling of the conjunctiva. The skin may also be blue tinged because of lack of oxygen.
Respiratory
Respiratory symptoms and signs that may be present include shortness of breath, wheezes, or stridor. The wheezing is typically caused by spasms of the bronchial muscles while stridor is related to upper airway obstruction secondary to swelling. Hoarseness, pain with swallowing, or a cough may also occur.
Cardiovascular
While a fast heart rate caused by low blood pressure is more common, a Bezold–Jarisch reflex has been described in 10% of people, where a slow heart rate is associated with low blood pressure. A drop in blood pressure or shock (either distributive or cardiogenic) may cause the feeling of lightheadedness or loss of consciousness. Rarely very low blood pressure may be the only sign of anaphylaxis.
Coronary artery spasm may occur with subsequent myocardial infarction, dysrhythmia, or cardiac arrest. Those with underlying coronary disease are at greater risk of cardiac effects from anaphylaxis. The coronary spasm is related to the presence of histamine-releasing cells in the heart.
Other
Gastrointestinal symptoms may include severe crampy abdominal pain, diarrhea, and vomiting. There may be confusion, a loss of bladder control or pelvic pain similar to that of uterine cramps. Dilation of blood vessels around the brain may cause headaches. A feeling of anxiety or of "impending doom" has also been described.
Causes
Anaphylaxis can occur in response to almost any foreign substance. Common triggers include venom from insect bites or stings, foods, and medication. Foods are the most common trigger in children and young adults, while medications and insect bites and stings are more common in older adults. Less common causes include: physical factors, biological agents such as semen, latex, hormonal changes, food additives such as monosodium glutamate and food colors, and topical medications. Physical factors such as exercise (known as exercise-induced anaphylaxis) or temperature (either hot or cold) may also act as triggers through their direct effects on mast cells. Events caused by exercise are frequently associated with cofactors such as the ingestion of certain foods or taking an NSAID. During anesthesia, neuromuscular blocking agents, antibiotics, and latex are the most common causes. The cause remains unknown in 32–50% of cases, referred to as "idiopathic anaphylaxis." Six vaccines (MMR, varicella, influenza, hepatitis B, tetanus, meningococcal) are recognized as a cause for anaphylaxis, and HPV may cause anaphylaxis as well.
Food
Many foods can trigger anaphylaxis; this may occur upon the first known ingestion. Common triggering foods vary around the world due to cultural cuisine. In Western cultures, ingestion of or exposure to peanuts, wheat, nuts, certain types of seafood like shellfish, milk, fruit and eggs are the most prevalent causes. Sesame is common in the Middle East, while rice and chickpeas are frequently encountered as sources of anaphylaxis in Asia. Severe cases are usually caused by ingesting the allergen, but some people experience a severe reaction upon contact. Children can outgrow their allergies. By age 16, 80% of children with anaphylaxis to milk or eggs and 20% who experience isolated anaphylaxis to peanuts can tolerate these foods.
Medication
Any medication may potentially trigger anaphylaxis. The most common are β-lactam antibiotics (such as penicillin) followed by aspirin and NSAIDs. Other antibiotics are implicated less frequently. Anaphylactic reactions to NSAIDs are either agent specific or occur among those that are structurally similar meaning that those who are allergic to one NSAID can typically tolerate a different one or different group of NSAIDs. Other relatively common causes include chemotherapy, vaccines, protamine and herbal preparations. Some medications (vancomycin, morphine, x-ray contrast among others) cause anaphylaxis by directly triggering mast cell degranulation.
The frequency of a reaction to an agent partly depends on the frequency of its use and partly on its intrinsic properties. Anaphylaxis to penicillin or cephalosporins occurs only after it binds to proteins inside the body with some agents binding more easily than others. Anaphylaxis to penicillin occurs once in every 2,000 to 10,000 courses of treatment, with death occurring in fewer than one in every 50,000 courses of treatment. Anaphylaxis to aspirin and NSAIDs occurs in about one in every 50,000 persons. If someone has a reaction to penicillin, his or her risk of a reaction to cephalosporins is greater but still less than one in 1,000. The old radiocontrast agents caused reactions in 1% of cases, while the newer lower osmolar agents cause reactions in 0.04% of cases.
Venom
Venom from stinging or biting insects such as Hymenoptera (ants, bees, and wasps) or Triatominae (kissing bugs) may cause anaphylaxis in susceptible people. Previous reactions that are anything more than a local reaction around the site of the sting, are a risk factor for future anaphylaxis; however, half of fatalities have had no previous systemic reaction.
Risk factors
People with atopic diseases such as asthma, eczema, or allergic rhinitis are at high risk of anaphylaxis from food, latex, and radiocontrast agents but not from injectable medications or stings. One study in children found that 60% had a history of previous atopic diseases, and of children who die from anaphylaxis, more than 90% have asthma. Those with mastocytosis or of a higher socioeconomic status are at increased risk.
Pathophysiology
Anaphylaxis is a severe allergic reaction of rapid onset affecting many body systems. It is due to the release of inflammatory mediators and cytokines from mast cells and basophils, typically due to an immunologic reaction but sometimes non-immunologic mechanism.
Interleukin (IL)–4 and IL-13 are cytokines important in the initial generation of antibody and inflammatory cell responses to anaphylaxis.
Immunologic
In the immunologic mechanism, immunoglobulin E (IgE) binds to the antigen (the foreign material that provokes the allergic reaction). Antigen-bound IgE then activates FcεRI receptors on mast cells and basophils. This leads to the release of inflammatory mediators such as histamine. These mediators subsequently increase the contraction of bronchial smooth muscles, trigger vasodilation, increase the leakage of fluid from blood vessels, and cause heart muscle depression. There is also a non-immunologic mechanism that does not rely on IgE, but it is not known if this occurs in humans.
Non-immunologic
Non-immunologic mechanisms involve substances that directly cause the degranulation of mast cells and basophils. These include agents such as contrast medium, opioids, temperature (hot or cold), and vibration. Sulfites may cause reactions by both immunologic and non-immunologic mechanisms.
Diagnosis
Anaphylaxis is diagnosed on the basis of a person's signs and symptoms. When any one of the following three occurs within minutes or hours of exposure to an allergen there is a high likelihood of anaphylaxis:
Involvement of the skin or mucosal tissue plus either respiratory difficulty or a low blood pressure causing symptoms
Two or more of the following symptoms after a likely contact with an allergen:
a. Involvement of the skin or mucosa
b. Respiratory difficulties
c. Low blood pressure
d. Gastrointestinal symptoms
Low blood pressure after exposure to a known allergen
Skin involvement may include: hives, itchiness or a swollen tongue among others. Respiratory difficulties may include: shortness of breath, stridor, or low oxygen levels among others. Low blood pressure is defined as a greater than 30% decrease from a person's usual blood pressure. In adults a systolic blood pressure of less than 90 mmHg is often used.
During an attack, blood tests for tryptase or histamine (released from mast cells) might be useful in diagnosing anaphylaxis due to insect stings or medications. However these tests are of limited use if the cause is food or if the person has a normal blood pressure, and they are not specific for the diagnosis.
Classification
There are three main classifications of anaphylaxis.
Anaphylactic shock is associated with systemic vasodilation that causes low blood pressure which is by definition 30% lower than the person's baseline or below standard values.
Biphasic anaphylaxis is the recurrence of symptoms within 1–72 hours after resolution of an initial anaphylactic episode. Estimates of incidence vary, between less than 1% and up to 20% of cases. The recurrence typically occurs within 8 hours. It is managed in the same manner as anaphylaxis.
Anaphylactoid reaction, non-immune anaphylaxis, or pseudoanaphylaxis, is a type of anaphylaxis that does not involve an allergic reaction but is due to direct mast cell degranulation. Non-immune anaphylaxis is the current term, as of 2018, used by the World Allergy Organization with some recommending that the old terminology, "anaphylactoid", no longer be used.
Allergy testing
Allergy testing may help in determining the trigger. Skin allergy testing is available for certain foods and venoms. Blood testing for specific IgE can be useful to confirm milk, egg, peanut, tree nut and fish allergies.
Skin testing is available to confirm penicillin allergies, but is not available for other medications. Non-immune forms of anaphylaxis can only be determined by history or exposure to the allergen in question, and not by skin or blood testing.
Differential diagnosis
It can sometimes be difficult to distinguish anaphylaxis from asthma, syncope, and panic attacks. Asthma however typically does not entail itching or gastrointestinal symptoms, syncope presents with pallor rather than a rash, and a panic attack may have flushing but does not have hives. Other conditions that may present similarly include: scrombroidosis and anisakiasis.
Post-mortem findings
In a person who died from anaphylaxis, autopsy may show an "empty heart" attributed to reduced venous return from vasodilation and redistribution of intravascular volume from the central to the peripheral compartment. Other signs are laryngeal edema, eosinophilia in lungs, heart and tissues, and evidence of myocardial hypoperfusion. Laboratory findings could detect increased levels of serum tryptase, increase in total and specific IgE serum levels.
Prevention
Avoidance of the trigger of anaphylaxis is recommended. In cases where this may not be possible, desensitization may be an option. Immunotherapy with Hymenoptera venoms is effective at desensitizing 80–90% of adults and 98% of children against allergies to bees, wasps, hornets, yellowjackets, and fire ants. Oral immunotherapy may be effective at desensitizing some people to certain food including milk, eggs, nuts and peanuts; however, adverse effects are common. For example, many people develop an itchy throat, cough, or lip swelling during immunotherapy. Desensitization is also possible for many medications, however it is advised that most people simply avoid the agent in question. In those who react to latex it may be important to avoid cross-reactive foods such as avocados, bananas, and potatoes among others.
Management
Anaphylaxis is a medical emergency that may require resuscitation measures such as airway management, supplemental oxygen, large volumes of intravenous fluids, and close monitoring. Passive leg raise may also be helpful in the emergency management.
Administration of epinephrine is the treatment of choice with antihistamines and steroids (for example, dexamethasone) often used as adjuncts. A period of in-hospital observation for between 2 and 24 hours is recommended for people once they have returned to normal due to concerns of biphasic anaphylaxis.
Epinephrine
Epinephrine (adrenaline) (1 in 1,000) is the primary treatment for anaphylaxis with no absolute contraindication to its use. It is recommended that an epinephrine solution be given intramuscularly into the mid anterolateral thigh as soon as the diagnosis is suspected. The injection may be repeated every 5 to 15 minutes if there is insufficient response. A second dose is needed in 16–35% of episodes with more than two doses rarely required. The intramuscular route is preferred over subcutaneous administration because the latter may have delayed absorption. It is recommended that after diagnosis and treatment of anaphylaxis, the patient should be kept under observation in an appropriate clinical setting until symptoms have fully resolved. Minor adverse effects from epinephrine include tremors, anxiety, headaches, and palpitations.
People on β-blockers may be resistant to the effects of epinephrine. In this situation if epinephrine is not effective intravenous glucagon can be administered which has a mechanism of action independent of β-receptors.
If necessary, it can also be given intravenously using a dilute epinephrine solution. Intravenous epinephrine, however, has been associated both with dysrhythmia and myocardial infarction. Epinephrine autoinjectors used for self-administration typically come in two doses, one for adults or children who weigh more than 25 kg and one for children who weigh 10 to 25 kg.
Adjuncts
Antihistamines (both H1 and H2), while commonly used and assumed effective based on theoretical reasoning, are poorly supported by evidence. A 2007 Cochrane review did not find any good-quality studies upon which to base recommendations and they are not believed to have an effect on airway edema or spasm. Corticosteroids are unlikely to make a difference in the current episode of anaphylaxis, but may be used in the hope of decreasing the risk of biphasic anaphylaxis. Their prophylactic effectiveness in these situations is uncertain. Nebulized salbutamol may be effective for bronchospasm that does not resolve with epinephrine. Methylene blue has been used in those not responsive to other measures due to its presumed effect of relaxing smooth muscle.
Preparedness
People prone to anaphylaxis are advised to have an allergy action plan. Parents are advised to inform schools of their children's allergies and what to do in case of an anaphylactic emergency. The action plan usually includes use of epinephrine autoinjectors, the recommendation to wear a medical alert bracelet, and counseling on avoidance of triggers. Immunotherapy is available for certain triggers to prevent future episodes of anaphylaxis. A multi-year course of subcutaneous desensitization has been found effective against stinging insects, while oral desensitization is effective for many foods.
Prognosis
In those in whom the cause is known and prompt treatment is available, the prognosis is good. Even if the cause is unknown, if appropriate preventive medication is available, the prognosis is generally good. If death occurs, it is usually due to either respiratory (typically asphyxia) or cardiovascular causes (shock), with 0.7–20% of cases causing death. There have been cases of death occurring within minutes. Outcomes in those with exercise-induced anaphylaxis are typically good, with fewer and less severe episodes as people get older.
Epidemiology
The number of people who get anaphylaxis is 4–100 per 100,000 persons per year, with a lifetime risk of 0.05–2%. About 30% of people get more than one attack. Exercise-induced anaphylaxis affects about 1 in 2000 young people.
Rates appear to be increasing: the numbers in the 1980s were approximately 20 per 100,000 per year, while in the 1990s it was 50 per 100,000 per year. The increase appears to be primarily for food-induced anaphylaxis. The risk is greatest in young people and females.
Anaphylaxis leads to as many as 500–1,000 deaths per year (2.7 per million) in the United States, 20 deaths per year in the United Kingdom (0.33 per million), and 15 deaths per year in Australia (0.64 per million). Another estimate from the United States puts the death rate at 0.7 per million. Mortality rates have decreased between the 1970s and 2000s. In Australia, death from food-induced anaphylaxis occur primarily in women while deaths due to insect bites primarily occur in males. Death from anaphylaxis is most commonly triggered by medications.
History
The term aphylaxis was coined by French physiologist Charles Richet in 1902 in the sense "lack of protection". Richet himself later changed the term to anaphylaxis on grounds of euphony. The term is from the Greek ἀνά-, ana-, meaning "against", and φύλαξις, phylaxis, meaning "protection".
In his experiments, Richet injected a dog with sea anemone (Actinia) toxin in an attempt to protect it. Although the dog had previously tolerated the toxin, on re-exposure, three weeks later with the same dose, it developed fatal anaphylaxis. Thus instead of inducing tolerance (prophylaxis), when lethal responses resulted from previously tolerated doses, he coined the word a (without) phylaxis (protection). He was subsequently awarded the Nobel Prize in Physiology or Medicine for his work on anaphylaxis in 1913. The phenomenon itself, however, has been described since ancient times.
Research
There are ongoing efforts to develop sublingual epinephrine to treat anaphylaxis. Trials of sublingual epinephrine, currently called AQST-108 (dipivefrin) and sponsored by Aquestive Therapeutics, are in phase 1 trials as of December 2021. Subcutaneous injection of the anti-IgE antibody omalizumab is being studied as a method of preventing recurrence, but it is not yet recommended.
References
External links
and Anaphylaxis pathway
Allergology
Complications of surgical and medical care
Medical emergencies
Respiratory diseases
Type I hypersensitivity
Urticaria and angioedema
Wikipedia medicine articles ready to translate (full)
Wikipedia emergency medicine articles ready to translate |
null | null | Poker probability | eng_Latn | In poker, the probability of each type of 5-card hand can be computed by calculating the proportion of hands of that type among all possible hands.
History
Probability and gambling have been ideas since long before the invention of poker. The development of probability theory in the late 1400s was attributed to gambling; when playing a game with high stakes, players wanted to know what the chance of winning would be. In 1494, Fra Luca Paccioli released his work which was the first written text on probability. Motivated by Paccioli's work, Girolamo Cardano (1501-1576) made further developments in probability theory. His work from 1550, titled Liber de Ludo Aleae, discussed the concepts of probability and how they were directly related to gambling. However, his work did not receive any immediate recognition since it was not published until after his death. Blaise Pascal (1623-1662) also contributed to probability theory. His friend, Chevalier de Méré, was an avid gambler with the goal to become wealthy from it. De Méré tried a new mathematical approach to a gambling game but did not get the desired results. Determined to know why his strategy was unsuccessful, he consulted with Pascal. Pascal's work on this problem began an important correspondence between him and fellow mathematician Pierre de Fermat (1601-1665). Communicating through letters, the two continued to exchange their ideas and thoughts. These interactions led to the conception of basic probability theory. To this day, many gamblers still rely on the basic concepts of probability theory in order to make informed decisions while gambling.
Frequency of 5-card poker hands
The following chart enumerates the (absolute) frequency of each hand, given all combinations of 5 cards randomly drawn from a full deck of 52 without replacement. Wild cards are not considered. In this chart:
Distinct hands is the number of different ways to draw the hand, not counting different suits.
Frequency is the number of ways to draw the hand, including the same card values in different suits.
The Probability of drawing a given hand is calculated by dividing the number of ways of drawing the hand (Frequency) by the total number of 5-card hands (the sample space; ). For example, there are 4 different ways to draw a royal flush (one for each suit), so the probability is , or one in 649,740. One would then expect to draw this hand about once in every 649,740 draws, or nearly 0.000154% of the time.
Cumulative probability refers to the probability of drawing a hand as good as or better than the specified one. For example, the probability of drawing three of a kind is approximately 2.11%, while the probability of drawing a hand at least as good as three of a kind is about 2.87%. The cumulative probability is determined by adding one hand's probability with the probabilities of all hands above it.
The Odds are defined as the ratio of the number of ways not to draw the hand, to the number of ways to draw it. In statistics, this is called odds against. For instance, with a royal flush, there are 4 ways to draw one, and 2,598,956 ways to draw something else, so the odds against drawing a royal flush are 2,598,956 : 4, or 649,739 : 1. The formula for establishing the odds can also be stated as (1/p) - 1 : 1, where p is the aforementioned probability.
The values given for Probability, Cumulative probability, and Odds are rounded off for simplicity; the Distinct hands and Frequency values are exact.
The nCr function on most scientific calculators can be used to calculate hand frequencies; entering with and , for example, yields as above.
The royal flush is a case of the straight flush. It can be formed 4 ways (one for each suit), giving it a probability of 0.000154% and odds of 649,739 : 1.
When ace-low straights and ace-low straight flushes are not counted, the probabilities of each are reduced: straights and straight flushes each become 9/10 as common as they otherwise would be. The 4 missed straight flushes become flushes and the 1,020 missed straights become no pair.
Note that since suits have no relative value in poker, two hands can be considered identical if one hand can be transformed into the other by swapping suits. For example, the hand 3♣ 7♣ 8♣ Q♠ A♠ is identical to 3♦ 7♦ 8♦ Q♥ A♥ because replacing all of the clubs in the first hand with diamonds and all of the spades with hearts produces the second hand. So eliminating identical hands that ignore relative suit values, there are only 134,459 distinct hands.
The number of distinct poker hands is even smaller. For example, 3♣ 7♣ 8♣ Q♠ A♠ and 3♦ 7♣ 8♦ Q♥ A♥ are not identical hands when just ignoring suit assignments because one hand has three suits, while the other hand has only two—that difference could affect the relative value of each hand when there are more cards to come. However, even though the hands are not identical from that perspective, they still form equivalent poker hands because each hand is an A-Q-8-7-3 high card hand. There are 7,462 distinct poker hands.
Frequency of 7-card poker hands
In some popular variations of poker such as Texas hold 'em, a player uses the best five-card poker hand out of seven cards. The frequencies are calculated in a manner similar to that shown for 5-card hands, except additional complications arise due to the extra two cards in the 7-card poker hand. The total number of distinct 7-card hands is . It is notable that the probability of a no-pair hand is less than the probability of a one-pair or two-pair hand.
The Ace-high straight flush or royal flush is slightly more frequent (4324) than the lower straight flushes (4140 each) because the remaining two cards can have any value; a King-high straight flush, for example, cannot have the Ace of its suit in the hand (as that would make it ace-high instead).
{| class="wikitable" style="font-size: 95%; text-align:center;"
|-
!Hand !! Frequency !! Probability !! Cumulative !! Odds against
!Mathematical expression of absolute frequency
|-
|Royal flush
|align=center| 4,324 || 0.0032% || 0.0032% ||align=center| 30,939 : 1
|
|-
|Straight flush (excluding royal flush)
|align=center| 37,260 || 0.0279% || 0.0311% ||align=center| 3,589.6 : 1
|
|-
|Four of a kind
|align=center| 224,848 || 0.168% || 0.199% ||align=center| 594 : 1
|
|-
|Full house
|align=center| 3,473,184 || 2.60% || 2.80% ||align=center| 37.5 : 1
|
|-
|Flush (excluding royal flush and straight flush)
|align=center| 4,047,644 || 3.03% || 5.82% ||align=center| 32.1 : 1
|
|-
|Straight (excluding royal flush and straight flush)
|align=center| 6,180,020 || 4.62% || 10.4% ||align=center| 20.6 : 1
|
|-
|Three of a kind
|align=center| 6,461,620 || 4.83% || 15.3% ||align=center| 19.7 : 1
|
|-
|Two pair
|align=center| 31,433,400 || 23.5% || 38.8% ||align=center| 3.26 : 1
|
|-
|One pair
|align=center| 58,627,800 || 43.8% || 82.6% ||align=center| 1.28 : 1
|
|-
|No pair / High card
|align=center| 23,294,460 || 17.4% || 100% ||align=center| 4.74 : 1
|
|-
! Total
! align="center" | 133,784,560 !! 100% !! --- !! align="center" | 0 : 1
!
|}
(The frequencies given are exact; the probabilities and odds are approximate.)
Since suits have no relative value in poker, two hands can be considered identical if one hand can be transformed into the other by swapping suits. Eliminating identical hands that ignore relative suit values leaves 6,009,159 distinct 7-card hands.
The number of distinct 5-card poker hands that are possible from 7 cards is 4,824. Perhaps surprisingly, this is fewer than the number of 5-card poker hands from 5 cards because some 5-card hands are impossible with 7 cards (e.g. 7-high).
Frequency of 5-card lowball poker hands
Some variants of poker, called lowball, use a low hand to determine the winning hand. In most variants of lowball, the ace is counted as the lowest card and straights and flushes don't count against a low hand, so the lowest hand is the five-high hand A-2-3-4-5, also called a wheel. The probability is calculated based on , the total number of 5-card combinations. (The frequencies given are exact; the probabilities and odds are approximate.)
{| class="wikitable" style="font-size: 95%;"
|-
!Hand !! Distinct hands !! Frequency !! Probability !! Cumulative !! Odds against
|-
| 5-high ||align=right| 1 ||align=right| 1,024 || 0.0394% || 0.0394% ||align=right| 2,537.05 : 1
|-
| 6-high ||align=right| 5 ||align=right| 5,120 || 0.197% || 0.236% ||align=right| 506.61 : 1
|-
| 7-high ||align=right| 15 ||align=right| 15,360 || 0.591% || 0.827% ||align=right| 168.20 : 1
|-
| 8-high ||align=right| 35 ||align=right| 35,840 || 1.38% || 2.21% ||align=right| 71.52 : 1
|-
| 9-high ||align=right| 70 ||align=right| 71,680 || 2.76% || 4.96% ||align=right| 35.26 : 1
|-
| 10-high ||align=right| 126 ||align=right| 129,024 || 4.96% || 9.93% ||align=right| 19.14 : 1
|-
| Jack-high ||align=right| 210 ||align=right| 215,040 || 8.27% || 18.2% ||align=right| 11.09 : 1
|-
| Queen-high ||align=right| 330 ||align=right| 337,920 || 13.0% || 31.2% ||align=right| 6.69 : 1
|-
| King-high ||align=right| 495 ||align=right| 506,880 || 19.5% || 50.7% ||align=right| 4.13 : 1
|-
|Total ||align=right|1,287 ||align=right|1,317,888 || 50.7% || 50.7% ||align=right| 0.97 : 1
|}
As can be seen from the table, just over half the time a player gets a hand that has no pairs, threes- or fours-of-a-kind. (50.7%)
If aces are not low, simply rotate the hand descriptions so that 6-high replaces 5-high for the best hand and ace-high replaces king-high as the worst hand.
Some players do not ignore straights and flushes when computing the low hand in lowball. In this case, the lowest hand is A-2-3-4-6 with at least two suits. Probabilities are adjusted in the above table such that "5-high" is not listed", "6-high" has one distinct hand, and "King-high" having 330 distinct hands, respectively. The Total line also needs adjusting.
Frequency of 7-card lowball poker hands
In some variants of poker a player uses the best five-card low hand selected from seven cards. In most variants of lowball, the ace is counted as the lowest card and straights and flushes don't count against a low hand, so the lowest hand is the five-high hand A-2-3-4-5, also called a wheel. The probability is calculated based on , the total number of 7-card combinations.
The table does not extend to include five-card hands with at least one pair. Its "Total" represents the 95.4% of the time that a player can select a 5-card low hand without any pair.
{| class="wikitable" style="font-size: 95%;"
|-
!Hand !! Frequency !! Probability !! Cumulative !! Odds against
|-
| 5-high ||align=right| 781,824 || 0.584% || 0.584% ||align=right| 170.12 : 1
|-
| 6-high ||align=right| 3,151,360 || 2.36% || 2.94% ||align=right| 41.45 : 1
|-
| 7-high ||align=right| 7,426,560 || 5.55% || 8.49% ||align=right| 17.01 : 1
|-
| 8-high ||align=right| 13,171,200 || 9.85% || 18.3% ||align=right| 9.16 : 1
|-
| 9-high ||align=right| 19,174,400 || 14.3% || 32.7% ||align=right| 5.98 : 1
|-
| 10-high ||align=right| 23,675,904 || 17.7% || 50.4% ||align=right| 4.65 : 1
|-
| Jack-high ||align=right| 24,837,120 || 18.6% || 68.9% ||align=right| 4.39 : 1
|-
| Queen-high ||align=right| 21,457,920 || 16.0% || 85.0% ||align=right| 5.23 : 1
|-
| King-high ||align=right| 13,939,200 || 10.4% || 95.4% ||align=right| 8.60 : 1
|-
|Total ||align=right| 127,615,488 || 95.4% || 95.4% ||align=right| 0.05 : 1
|}
(The frequencies given are exact; the probabilities and odds are approximate.)
If aces are not low, simply rotate the hand descriptions so that 6-high replaces 5-high for the best hand and ace-high replaces king-high as the worst hand.
Some players do not ignore straights and flushes when computing the low hand in lowball. In this case, the lowest hand is A-2-3-4-6 with at least two suits. Probabilities are adjusted in the above table such that "5-high" is not listed, "6-high" has 781,824 distinct hands, and "King-high" has 21,457,920 distinct hands, respectively. The Total line also needs adjusting.
See also
Probability
Odds
Sample space
Event (probability theory)
Binomial coefficient
Combination
Permutation
Combinatorial game theory
Game complexity
Set theory
Gaming mathematics
Notes
External links
Brian Alspach's mathematics and poker page
MathWorld: Poker
Poker probabilities including conditional calculations
Numerous poker probability tables
5, 6, and 7 card poker probabilities
Poker Odds for Dummies
The 7,462 and 4,824 equivalence classes
Preflop, After Flop and Chance of Making Hand Odds
Odds and Outs probability table
Poker probability calculator 5, 6 and 7 cards
Visual odds calculator |
null | null | List of Interstate Highways | eng_Latn | There are 70 primary Interstate Highways in the Interstate Highway System, a network of freeways in the United States. They are assigned one- or two-digit route numbers, whereas their associated "auxiliary" Interstate Highways receive three-digit route numbers. Typically, odd-numbered Interstates run south-north, with lower numbers in the west and higher numbers in the east; even-numbered Interstates run west-east, with lower numbers in the south and higher numbers in the north. Highways whose route numbers are divisible by 5 usually represent major coast-to-coast or border-to-border routes (ex. I-10 travels from Santa Monica, California, to Jacksonville, Florida, traveling from the Pacific to Atlantic oceans). Additionally, auxiliary highways have their numbering system where a different number prefixes the number of its parent highway.
Five route numbers are duplicated in the system, though the corresponding highways are separated by state lines which prevent confusion. The main list that discusses the primary Interstate Highways in the contiguous United States is followed by sections regarding Alaska, Hawaii, and Puerto Rico.
Contiguous United States
There are 70 primary Interstate Highways listed in the table below.
Other jurisdictions
In addition to the 48 contiguous states, Interstate Highways are found in Hawaii, Alaska, and Puerto Rico. The Federal Highway Administration funds four routes in Alaska and three routes in Puerto Rico under the same program as the rest of the Interstate Highway System. However, these routes are not required to meet the same standards as the mainland routes:
Hawaii
The Interstate Highways in the island of Oahu, Hawaii are signed with the standard Interstate Highway shield, with the letter "H-" prefixed before the number. They are fully controlled-access routes built to the same standards as the mainland Interstate Highways.
Alaska
Alaska's Interstate Highways are unsigned as such, although they all have state highway numbers that do not match the Interstate Highway numbers.
Puerto Rico
Like Alaska, Puerto Rico signs its Interstate Highways as territorial routes, and the numbers do not match their official Interstate Highway designations. Many of the territory's routes are freeway-standard toll roads.
In popular culture
The title of interactive fiction I-0 refers to the fictional "Interstate Zero". A fictional highway is depicted in the movie Interstate 60.
See also
References
External links
Alaska Roads - Interstate ends photos
Puerto Rico road photos
The Dwight D. Eisenhower National System of Interstate and Defense Highways Route Log and Finder List
3-digit Interstate Highways
Interstate |
null | null | Movie theater | eng_Latn | A movie theater (American English), cinema (British English), or cinema hall (Indian English), also known as a picture house, the pictures, picture theater, the silver screen, or the movies, is a building that contains auditoria for viewing films (also called movies) for entertainment. Most, but not all, theaters are commercial operations catering to the general public, who attend by purchasing a ticket. Some movie theaters, however, are operated by non-profit organizations or societies that charge members a membership fee to view films.
The film is projected with a movie projector onto a large projection screen at the front of the auditorium while the dialogue, sounds, and music are played through a number of wall-mounted speakers. Since the 1970s, subwoofers have been used for low-pitched sounds. Since the 2010s, most movie theaters have been equipped for digital cinema projection, removing the need to create and transport a physical film print on a heavy reel.
A great variety of films are shown at cinemas, ranging from animated films to blockbusters to documentaries. The smallest movie theaters have a single viewing room with a single screen. In the 2010s, most movie theaters had multiple screens. The largest theater complexes, which are called multiplexes—a concept developed in Canada in the 1950s — have up to thirty screens. The audience members often sit on padded seats, which in most theaters are set on a sloped floor, with the highest part at the rear of the theater. Movie theaters often sell soft drinks, popcorn, and candy, and some theaters sell hot fast food. In some jurisdictions, movie theaters can be licensed to sell alcoholic drinks.
Terminology
A movie theater may also be referred to as a movie house, film house, film theater, cinema or picture house. In the US, theater has long been the preferred spelling, while in the UK, Australia, Canada and elsewhere it's theatre.
However, some US theaters opt to use the British spelling in their own names, a practice supported by the National Association of Theatre Owners, while apart from Anglophone North America most English-speaking countries use the term cinema , alternatively spelled and pronounced kinema . The latter terms, as well as their derivative adjectives "cinematic" and "kinematic", ultimately derive from Greek κινῆμα, κινήματος (kinema, kinematos)—"movement", "motion". In the countries where those terms are used, the word "theatre" is usually reserved for live performance venues.
Colloquial expressions, mostly applied to motion pictures and motion picture theaters collectively, include the silver screen (formerly sometimes sheet) and the big screen (contrasted with the smaller screen of a television set). Specific to North American term is the movies, while specific terms in the UK are the pictures, the flicks and for the facility itself the flea pit (or fleapit). A screening room is a small theater, often a private one, such as for the use of those involved in the production of motion pictures or in a large private residence.
The etymology of the term "movie theater" involves the term "movie", which is a "shortened form of moving picture in the cinematographic sense" that was first used in 1896 and "theater", which originated in the "...late 14c., [meaning an] open air place in ancient times for viewing spectacles and plays". The term "theater" comes from the Old French word "theatre", from the 12th century and "...directly from Latin theatrum [which meant] 'play-house, theater; stage; spectators in a theater'", which in turn came from the Greek word "theatron", which meant "theater; the people in the theater; a show, a spectacle", [or] literally "place for viewing". The use of the word "theatre" to mean a "building where plays are shown" dates from the 1570s in the English language.
History
Precursors
Movie theatres stand in a long tradition of theaters that could house all kinds of entertainment. Some forms of theatrical entertainment would involve the screening of moving images and can be regarded as precursors of film.
In 1799, Étienne-Gaspard "Robertson" Robert moved his Phantasmagorie show to an abandoned cloister near the Place Vendôme in Paris. The eerie surroundings, with a graveyard and ruins, formed an ideal location for his ghostraising spectacle.
When it opened in 1838, The Royal Polytechnic Institution in London became a very popular and influential venue with all kinds of magic lantern shows as an important part of its program. At the main theatre, with 500 seats, lanternists would make good use of a battery of six large lanterns running on tracked tables to project the finely detailed images of extra large slides on the 648 square feet screen. The magic lantern was used to illustrate lectures, concerts, pantomimes and other forms of theatre. Popular magic lantern presentations included phantasmagoria, mechanical slides, Henry Langdon Childe's dissolving views and his chromatrope.
The earliest known public screening of projected stroboscopic animation was presented by Austrian magician Ludwig Döbler on 15 January 1847 at the Josephstadt Theatre in Vienna, with his patented Phantaskop. The animated spectacle was part of a well-received show that sold-out in several European cities during a tour that lasted until the spring of 1848.
The famous Parisian entertainment venue Le Chat Noir opened in 1881 and is remembered for its shadow plays, renewing the popularity of such shows in France.
Earliest motion picture screening venues
The earliest public film screenings took place in existing (vaudeville) theatres and other venues that could be darkened and comfortably house an audience.
Émile Reynaud screened his Pantomimes Lumineuses animated movies from 28 October 1892 to March 1900 at the Musée Grévin in Paris, with his Théâtre Optique system. He gave over 12,800 shows to a total of over 500,000 visitors, with programs including Pauvre Pierrot and Autour d'une cabine.
Thomas Edison initially believed film screening would not be as viable commercially as presenting films in peep boxes, hence the film apparatus that his company would first exploit became the kinetoscope. A few public demonstrations occurred since 9 May 1893, before a first public Kinetoscope parlor was opened on 14 April 1894, by the Holland Bros. in New York City at 1155 Broadway, on the corner of 27th Street. This can be regarded as the first commercial motion picture house. The venue had ten machines, set up in parallel rows of five, each showing a different movie. For 25 cents a viewer could see all the films in either row; half a dollar gave access to the entire bill.
The Eidoloscope, devised by Eugene Augustin Lauste for the Latham family, was demonstrated for members of the press on 21 April 1895 and opened to the paying public on 20 May, in a lower Broadway store with films of the Griffo-Barnett prize boxing fight, taken from Madison Square Garden's roof on 4 May.
Max Skladanowsky and his brother Emil demonstrated their motion pictures with the Bioscop in July 1895 at the Gasthaus Sello in Pankow (Berlin). This venue was later, at least since 1918, exploited as the full-time movie theatre Pankower Lichtspiele and between 1925 and 1994 as Tivoli. The first certain commercial screenings by the Skladanowsky brothers took place at the Wintergarten in Berlin from 1 to 31 November 1895.
The first commercial, public screening of films made with Louis and Auguste Lumière's Cinématographe took place in the basement of Salon Indien du Grand Café in Paris on 28 December 1895.
Early dedicated movie theatres
During the first decade of motion pictures, the demand for movies, the amount of new productions, and the average runtime of movies, all kept increasing, and at some stage it was viable to have theatres that would no longer program live acts, but only movies.
Claimants for the title of the earliest movie theatre include the Eden Theatre in La Ciotat, where L'Arrivée d'un train en gare de La Ciotat was screened on 21 March 1899. The theatre closed in 1995 but re-opened in 2013.
L'Idéal Cinéma in Aniche (France), built in 1901 as l’Hôtel du Syndicat CGT, showed its first film on 23 November 1905. The cinema was closed in 1977 and the building was demolished in 1993. The "Centre Culturel Claude Berri" was built in 1995; it integrates a new movie theater ( the Idéal Cinéma Jacques Tati).
In the United States, many small and simple theatres were set up, usually in converted storefronts. They typically charged five cents for admission, and thus became known as nickelodeons. This type of theatre flourished from about 1905 to circa 1915.
The Korsør Biograf Teater, in Korsør, Denmark, opened in August 1908 and is the oldest known movie theatre still in continuous operation.
Design
Traditionally a movie theater, like a stage theater, consists of a single auditorium with rows of comfortable padded seats, as well as a foyer area containing a box office for buying tickets. Movie theaters also often have a concession stand for buying snacks and drinks within the theater's lobby. Other features included are film posters, arcade games and washrooms. Stage theaters are sometimes converted into movie theaters by placing a screen in front of the stage and adding a projector; this conversion may be permanent, or temporary for purposes such as showing arthouse fare to an audience accustomed to plays. The familiar characteristics of relatively low admission and open seating can be traced to Samuel Roxy Rothafel, an early movie theater impresario. Many of these early theaters contain a balcony, an elevated level across the auditorium above the theater's rearmost seats. The rearward main floor "loge" seats were sometimes larger, softer, and more widely spaced and sold for a higher price. In conventional low pitch viewing floors the preferred seating arrangement is to use staggered rows. While a less efficient use of floor space this allows a somewhat improved sight line between the patrons seated in the next row toward the screen, provided they do not lean toward one another.
"Stadium seating", popular in modern multiplexes, actually dates back to the 1920s. The 1922 Princess Theatre in Honolulu, Hawaii featured "stadium seating", sharply raked rows of seats extending from in front of the screen back towards the ceiling. It gives patrons a clear sight line over the heads of those seated in front of them. Modern "stadium seating" was utilized in IMAX theaters, which have very tall screens, beginning in the early 1970s.
Rows of seats are divided by one or more aisles so that there are seldom more than 20 seats in a row. This allows easier access to seating, as the space between rows is very narrow. Depending on the angle of rake of the seats, the aisles have steps. In older theaters, aisle lights were often built into the end seats of each row to help patrons find their way in the dark. Since the advent of stadium theaters with stepped aisles, each step in the aisles may be outlined with small lights to prevent patrons from tripping in the darkened theater. In movie theaters, the auditorium may also have lights that go to a low level, when the movie is going to begin. Theaters often have booster seats for children and other short people to put on the seat, to sit higher, for a better view. Many modern theaters have accessible seating areas for patrons in wheelchairs. See also luxury screens below.
Multiplexes and megaplexes
Canada was the first country in the world to have a two-screen theater. The Elgin Theatre in Ottawa, Ontario became the first venue to offer two film programs on different screens in 1957 when Canadian theater-owner Nat Taylor converted the dual screen theater into one capable of showing two different movies simultaneously. Taylor is credited by Canadian sources as the inventor of the multiplex or cineplex; he later founded the Cineplex Odeon Corporation, opening the 18-screen Toronto Eaton Centre Cineplex, the world's largest at the time, in Toronto, Ontario. In the United States, Stanley Durwood of American Multi-Cinema (now AMC Theatres) is credited as pioneering the multiplex in 1963 after realizing that he could operate several attached auditoriums with the same staff needed for one through careful management of the start times for each movie. Ward Parkway Center in Kansas City, Missouri had the first multiplex cinema in the United States.
Since the 1960s, multiple-screen theaters have become the norm, and many existing venues have been retrofitted so that they have multiple auditoriums. A single foyer area is shared among them. In the 1970s, many large 1920s movie palaces were converted into multiple screen venues by dividing their large auditoriums, and sometimes even the stage space, into smaller theaters. Because of their size, and amenities like plush seating and extensive food/beverage service, multiplexes and megaplexes draw from a larger geographic area than smaller theaters. As a rule of thumb, they pull audiences from an eight to 12-mile radius, versus a three to five-mile radius for smaller theaters (though the size of this radius depends on population density). As a result, the customer geography area of multiplexes and megaplexes typically overlaps with smaller theaters, which face threat of having their audience siphoned by bigger theaters that cut a wider swath in the movie-going landscape.
In most markets, nearly all single-screen theaters (sometimes referred to as a "Uniplex") have gone out of business; the ones remaining are generally used for arthouse films, e.g. the Crest Theatre in downtown Sacramento, California, small-scale productions, film festivals or other presentations. Because of the late development of multiplexes, the term "cinema" or "theater" may refer either to the whole complex or a single auditorium, and sometimes "screen" is used to refer to an auditorium. A popular film may be shown on multiple screens at the same multiplex, which reduces the choice of other films but offers more choice of viewing times or a greater number of seats to accommodate patrons. Two or three screens may be created by dividing up an existing cinema (as Durwood did with his Roxy in 1964), but newly built multiplexes usually have at least six to eight screens, and often as many as twelve, fourteen, sixteen or even eighteen.
Although definitions vary, a large multiplex with 20 or more screens is usually called a "megaplex". However, in the United Kingdom, this was a brand name for Virgin Cinema (later UGC). The first megaplex is generally considered to be the Kinepolis in Brussels, Belgium, which opened in 1988 with 25 screens and a seating capacity of 7,500. The first theater in the U.S. built from the ground up as a megaplex was the AMC Grand 24 in Dallas, Texas, which opened in May 1995, while the first megaplex in the U.S.-based on an expansion of an existing facility was Studio 28 in Grand Rapids, Michigan, which reopened in November 1988 with 20 screens and a seating capacity of 6,000.
Drive-in
A drive-in movie theater is an outdoor parking area with a screen—sometimes an inflatable screen—at one end and a projection booth at the other. Moviegoers drive into the parking spaces which are sometimes sloped upwards at the front to give a more direct view of the movie screen. Movies are usually viewed through the car windscreen (windshield) although some people prefer to sit on the bonnet (hood) of the car. Some may also sit in the trunk (back) of their car if space permits. Sound is either provided through portable loudspeakers located by each parking space, or is broadcast on an FM radio frequency, to be played through the car's stereo system. Because of their outdoor nature, drive-ins usually only operate seasonally, and after sunset. Drive-in movie theaters are mainly found in the United States, where they were especially popular in the 1950s and 1960s. Once numbering in the thousands, about 400 remain in the U.S. today. In some cases, multiplex or megaplex theaters were built on the sites of former drive-in theaters.
Other venues
Some outdoor movie theaters are just grassy areas where the audience sits upon chairs, blankets or even in hot tubs, and watch the movie on a temporary screen, or even the wall of a building. Colleges and universities have often sponsored movie screenings in lecture halls. The formats of these screenings include 35 mm, 16 mm, DVD, VHS, and even 70 mm in rare cases. Some alternative methods of showing movies have been popular in the past. In the 1980s the introduction of VHS cassettes made possible video-salons, small rooms where visitors viewed movies on a large TV. These establishments were especially popular in the Soviet Union, where official distribution companies were slow to adapt to changing demand, and so movie theaters could not show popular Hollywood and Asian films.
In 1967, the British government launched seven custom-built mobile cinema units for use as part of the Ministry of Technology campaign to raise standards. Using a very futuristic look, these 27-seat cinema vehicles were designed to attract attention. They were built on a Bedford SB3 chassis with a custom Coventry Steel Caravan extruded aluminum body. Movies are also commonly shown on airliners in flight, using large screens in each cabin or smaller screens for each group of rows or each individual seat; the airline company sometimes charges a fee for the headphones needed to hear the movie's sound. In a similar fashion, movies are sometimes also shown on trains, such as the Auto Train.
The smallest purpose-built cinema is the Cabiria Cine-Cafe which measures 24 m2 (258.3 ft²) and has a seating capacity of 18. It was built by Renata Carneiro Agostinho da Silva (Brazil) in Brasília DF, Brazil in 2008. It is mentioned in the 2010 Guinness World Records. The World's smallest solar-powered mobile cinema is Sol Cinema in the UK. Touring since 2010 the cinema is actually a converted 1972 caravan. It seats 8–10 at a time. In 2015 it featured in a Lenovo advert for the launch of a new tablet. The Bell Museum of Natural History in Minneapolis, Minnesota has recently begun summer "bike-ins", inviting only pedestrians or people on bicycles onto the grounds for both live music and movies. In various Canadian cities, including Toronto, Calgary, Ottawa and Halifax, al-fresco movies projected on the walls of buildings or temporarily erected screens in parks operate during the Summer and cater to a pedestrian audience. The New Parkway Museum in Oakland, California replaces general seating with couches and coffee tables, as well as having a full restaurant menu instead of general movie theater concessions such as popcorn or candy.
3D
3D film is a system of presenting film images so that they appear to the viewer to be three-dimensional. Visitors usually borrow or keep special glasses to wear while watching the movie. Depending on the system used, these are typically polarized glasses. Three-dimensional movies use two images channeled, respectively, to the right and left eyes to simulate depth by using 3-D glasses with red and blue lenses (anaglyph), polarized (linear and circular), and other techniques. 3-D glasses deliver the proper image to the proper eye and make the image appear to "pop-out" at the viewer and even follow the viewer when he/she moves so viewers relatively see the same image.
The earliest 3D movies were presented in the 1920s. There have been several prior "waves" of 3D movie distribution, most notably in the 1950s when they were promoted as a way to offer audiences something that they could not see at home on television. Still the process faded quickly and as yet has never been more than a periodic novelty in movie presentation. The "golden era" of 3D film began in the early 1950s with the release of the first color stereoscopic feature, Bwana Devil. The film starred Robert Stack, Barbara Britton and Nigel Bruce. James Mage was an early pioneer in the 3D craze. Using his 16 mm 3D Bolex system, he premiered his Triorama program in February 1953 with his four shorts: Sunday In Stereo, Indian Summer, American Life, and This is Bolex Stereo. 1953 saw two groundbreaking features in 3D: Columbia's Man in the Dark and Warner Bros. House of Wax, the first 3D feature with stereophonic sound. For many years, most 3-D movies were shown in amusement parks and even "4-D" techniques have been used when certain effects such as spraying of water, movement of seats, and other effects are used to simulate actions seen on the screen. The first decline in the theatrical 3D craze started in August and September 1953.
In 2009, movie exhibitors became more interested in 3D film. The number of 3D screens in theaters is increasing. The RealD company expects 15,000 screens worldwide in 2010. The availability of 3D movies encourages exhibitors to adopt digital cinema and provides a way for theaters to compete with home theaters. One incentive for theaters to show 3D films is that although ticket sales have declined, revenues from 3D tickets have grown. In the 2010s, 3D films became popular again. The IMAX 3D system and digital 3D systems are used (the latter is used in the animated movies of Disney/Pixar).
The RealD 3D system works by using a single digital projector that swaps back and forth between the images for eyes. A filter is placed in front of the projector that changes the polarization of the light coming from the projector. A silver screen is used to reflect this light back at the audience and reduce loss of brightness. There are four other systems available: Volfoni, Master Image, XpanD and Dolby 3D.
When a system is used that requires inexpensive 3D glasses, they can sometimes be kept by the patron. Most theaters have a fixed cost for 3D, while others charge for the glasses, but the latter is uncommon (at least in the United States). For example, in Pathé theaters in the Netherlands the extra fee for watching a 3D film consists of a fixed fee of €1.50, and an optional fee of €1 for the glasses. Holders of the Pathé Unlimited Gold pass (see also below) are supposed to bring along their own glasses; one pair, supplied yearly, more robust than the regular type, is included in the price.
IMAX
IMAX is a system using film with more than ten times the frame size of a 35 mm film to produce image quality far superior to conventional film. IMAX theaters use an oversized screen as well as special projectors. Invented by a Canadian company, the first permanent IMAX theater was at Ontario Place in Toronto, Canada. Until 2016, visitors to the IMAX cinema attached to the National Science and Media Museum in Bradford, West Yorkshire, England, United Kingdom, could observe the IMAX projection booth via a glass rear wall and watch the large format films being loaded and projected. There is also an IMAX theater in the Museum of science in Boston Massachusetts. The biggest movie theater screen in the world in Darling Harbour, Sydney Australia is an IMAX theater.
Programming
Movie theaters may be classified by the type of movies they show or when in a film's release process they are shown:
First-run theater: A theater that runs primarily mainstream film fare from the major film companies and distributors, during the initial new release period of each film.
Second-run or discount theater: A theater that runs films that have already shown in the first-run theaters and presented at a lower ticket price. (These are sometimes known as dollar theaters or "cheap seats".) This form of cinema is diminishing in viability owing to the increasingly shortened intervals before the films' home video release, called the "video window".
Repertoire/repertory theater or arthouse: A theater that presents more alternative and art films as well as second-run and classic films (often known as an "independent cinema" in the UK).
An adult movie theater or sex theater specializes in showing pornographic movies. Such movies are rarely shown in other theaters. See also Golden Age of Porn. Since the widespread availability of pornographic films for home viewing on VHS in the 1980s and 1990s, the DVD in the 1990s, and the Blu-ray disc in the 2000s, there are far fewer adult movie theaters.
IMAX theaters can show conventional movies, but the major benefits of the IMAX system are only available when showing movies filmed using it. While a few mainstream feature films have been produced in IMAX, IMAX movies are often documentaries featuring spectacular natural scenery, and may be limited to the 45-minute length of a single reel of IMAX film.
Presentation
Usually in the 2010s, an admission is for one feature film. Sometimes two feature films are sold as one admission (double feature), with a break in between. Separate admission for a short subject is rare; it is either an extra before a feature film or part of a series of short films sold as one admission (this mainly occurs at film festivals). (See also anthology film.) In the early decades of "talkie" films, many movie theaters presented a number of shorter items in addition to the feature film. This might include a newsreel, live-action comedy short films, documentary short films, musical short films, or cartoon shorts (many classic cartoons series such as the Looney Tunes and Mickey Mouse shorts were created for this purpose). Examples of this kind of programming are available on certain DVD releases of two of the most famous films starring Errol Flynn as a special feature arrangement designed to recreate that kind of filmgoing experience while the PBS series, Matinee at the Bijou, presented the equivalent content. Some theaters ran on continuous showings, where the same items would repeat throughout the day, with patrons arriving and departing at any time rather than having distinct entrance and exit cycles. Newsreels gradually became obsolete by the 1960s with the rise of television news, and most material now shown prior to a feature film is of a commercial or promotional nature (which usually include "trailers", which are advertisements for films and commercials for other consumer products or services).
A typical modern theater presents commercial advertising shorts, then movie trailers, and then the feature film. Advertised start times are usually for the entire program or session, not the feature itself; thus people who want to avoid commercials and trailers would opt to enter later. This is easiest and causes the least inconvenience when it is not crowded or one is not very choosy about where one wants to sit. If one has a ticket for a specific seat (see below) one is formally assured of that, but it is still inconvenient and disturbing to find and claim it during the commercials and trailers, unless it is near an aisle. Some movie theaters have some kind of break during the presentation, particularly for very long films. There may also be a break between the introductory material and the feature. Some countries such as the Netherlands have a tradition of incorporating an intermission in regular feature presentations, though many theaters have now abandoned that tradition, while in North America, this is very rare and usually limited to special circumstances involving extremely long movies. During the closing credits many people leave, but some stay until the end. Usually the lights are switched on after the credits, sometimes already during them. Some films show mid-credits scenes while the credits are rolling, which in comedy films are often bloopers and outtakes, or post-credits scenes, which typically set up the audience for a sequel.
Until the multiplex era, prior to showtime, the screen in some theaters would be covered by a curtain, in the style of a theater for a play. The curtain would be drawn for the feature. It is common practice in Australia for the curtain to cover part of the screen during advertising and trailers, then be fully drawn to reveal the full width of the screen for the main feature. Some theaters, lacking a curtain, filled the screen with slides of some form of abstract art prior to the start of the movie. Currently, in multiplexes, theater chains often feature a continuous slideshow between showings featuring a loop of movie trivia, promotional material for the theater chains (such as encouraging patrons to purchase drinks, snacks and popcorn, gift vouchers and group rates, or other foyer retail offers), or advertising for local and national businesses. Advertisements for Fandango and other convenient methods of purchasing tickets is often shown. Also prior to showing the film, reminders, in varying forms would be shown concerning theater etiquette (no smoking, no talking, no littering, removing crying babies, etc.) and in recent years, added reminders to silence mobile phones as well as warning concerning movie piracy with camcorders ("camming").
Some well-equipped theaters have "interlock" projectors which allow two or more projectors and sound units to be run in unison by connecting them electronically or mechanically. This set up can be used to project two prints in sync (for dual-projector 3-D) or to "interlock" one or more sound tracks to a single film. Sound interlocks were used for stereophonic sound systems before the advent of magnetic film prints. Fantasound (developed by RCA in 1940 for Disney's Fantasia) was an early interlock system. Likewise, early stereophonic films such as This Is Cinerama and House of Wax utilized a separate, magnetic oxide-coated film to reproduce up to six or more tracks of stereophonic sound. Datasat Digital Entertainment, purchaser of DTS's cinema division in May 2008, uses a time code printed on and read off of the film to synchronize with a CD-ROM in the sound track, allowing multi-channel soundtracks or foreign language tracks. This is not considered a projector interlock, however.
This practice is most common with blockbuster movies. Muvico Theaters, Regal Entertainment Group, Pacific Theatres and AMC Theatres are some theaters that interlock films.
Live broadcasting to movie theaters
Sometimes movie theaters provide digital projection of a live broadcast of an opera, concert, or other performance or event. For example, there are regular live broadcasts to movie theaters of Metropolitan Opera performances, with additionally limited repeat showings. Admission prices are often more than twice the regular movie theater admission prices.
Pricing and admission
In order to obtain admission to a movie theater, the prospective theater-goer must usually purchase a ticket from the box office, which may be for an arbitrary seat ("open" or "free" seating, first-come, first-served) or for a specific one (allocated seating). As of 2015, some theaters sell tickets online or at automated kiosks in the theater lobby. Movie theaters in North America generally have open seating. Cinemas in Europe can have free seating or numbered seating. Some theaters in Mexico offer numbered seating, in particular, Cinepolis VIP. In the case of numbered seating systems the attendee can often pick seats from a video screen. Sometimes the attendee cannot see the screen and has to make a choice based on a verbal description of the still available seats. In the case of free seats, already seated customers may be asked by staff to move one or more places for the benefit of an arriving couple or group wanting to sit together.
For 2013, the average price for a movie ticket in the United States was $8.13. The price of a ticket may be discounted during off-peak times e.g. for matinees, and higher at busy times, typically evenings and weekends. In Australia, Canada and New Zealand, when this practice is used, it is traditional to offer the lower prices for Tuesday for all showings, one of the slowest days of the week in the movie theater business, which has led to the nickname "cheap Tuesday". Sometimes tickets are cheaper on Monday, or on Sunday morning. Almost all movie theaters employ economic price discrimination: tickets for youth, students, and seniors are typically cheaper. Large theater chains, such as AMC Theaters, also own smaller theaters that show "second runs" of popular films, at reduced ticket prices. Movie theaters in India and other developing countries employ price discrimination in seating arrangement: seats closer to the screen cost less, while the ones farthest from the screen cost more. Movie theatres in India are also practicing safety guidelines and precautions after 2020.
In the United States, many movie theater chains sell discounted passes, which can be exchanged for tickets to regular showings. These passes are traditionally sold in bulk to institutional customers and also to the general public at Bulktix.com. Some passes provide substantial discounts from the price of regular admission, especially if they carry restrictions. Common restrictions include a waiting period after a movie's release before the pass can be exchanged for a ticket or specific theaters where a pass is ineligible for admission.
Some movie theaters and chains sell monthly passes for unlimited entrance to regular showings. Cinemas in Thailand have a restriction of one viewing per movie. The increasing number of 3D movies, for which an additional fee is required, somewhat undermines the concept of unlimited entrance to regular showings, in particular if no 2D version is screened, except in the cases where 3D is included. Some adult theaters sell a day pass, either as standard ticket, or as an option that costs a little more than a single admission. Also for some film festivals, a pass is sold for unlimited entrance. Discount theaters show films at a greatly discounted rate, however, the films shown are generally films that have already run for many weeks at regular theaters and thus are no longer a major draw, or films which flopped at the box office and thus have already been removed from showings at major theaters in order to free up screens for films that are a better box office draw.
Luxury screens
Some cinemas in city centers offer luxury seating with services like complimentary refills of soft drinks and popcorn, a bar serving beer, wine and liquor, reclining leather seats and service bells. Cinemas must have a liquor license to serve alcohol. The Vue Cinema and CGV Cinema chain is a good example of a large-scale offering of such a service, called "Gold Class" and similarly, ODEON, Britain's largest cinema chain, and 21 Cineplex, Indonesia's largest cinema chain, have gallery areas in some of their bigger cinemas where there is a separate foyer area with a bar and unlimited snacks.
Age restrictions
Admission to a movie may also be restricted by a motion picture rating system, typically due to depictions of sex, nudity or graphic violence. According to such systems, children or teenagers below a certain age may be forbidden access to theaters showing certain movies, or only admitted when accompanied by a parent or other adult. In some jurisdictions, a rating may legally impose these age restrictions on movie theaters. Where movie theaters do not have this legal obligation, they may enforce restrictions on their own. Accordingly, a movie theater may either not be allowed to program an unrated film, or voluntarily refrain from that.
Revenue
Movie studios/film distributors in the US traditionally drive hard bargains entitling them to as much as 100% of the gross ticket revenue during the first weeks (and then the balance changes in 10% increments in favor of exhibitors at intervals that vary from film to film). Film exhibition has seen a rise in its development with video consolidation as well as DVD sales, which over the past two decades is the biggest earner in revenue. According to The Contemporary Hollywood Film Industry, Philip Drake states that box office takings currently account for less than a quarter of total revenues and have become increasingly "front loaded", earning the majority of receipts in the opening two weeks of exhibition, meaning that films need to make an almost instant impact in order to avoid being dropped from screens by exhibitors. Essentially, if the film does not succeed in the first few weeks of its inception, it will most likely fail in its attempt to gain a sustainable amount of revenue and thus being taken out from movie theaters. Furthermore, higher-budget films on the "opening weekend", or the three days, Friday to Sunday, can signify how much revenue it will bring in, not only to America, but as well as overseas. It may also determine the price in distribution windows through home video and television.
In Canada, the total operating revenue in the movie theater industry was $1.7 billion in 2012, an 8.4% increase from 2010. This increase was mainly the result of growth in box office and concession revenue. Combined, these accounted for 91.9% of total industry operating revenue. In the US, the "...number of tickets sold fell nearly 11% between 2004 and 2013, according to the report, while box office revenue increased 17%" due to increased ticket prices.
New forms of competition
One reason for the decline in ticket sales in the 2000s is that "home-entertainment options [are] improving all the time— whether streamed movies and television, video games, or mobile apps—and studios releasing fewer movies", which means that "people are less likely to head to their local multiplex". This decline is not something that is recent. It has been observed since the 1950s when television became widespread among working-class homes. As the years went on, home media became more popular, and the decline continued. This decline continues until this day. A Pew Media survey from 2006 found that the relationship between movies watched at home versus at the movie theater was in a five to one ratio and 75% of respondents said their preferred way of watching a movie was at home, versus 21% who said they preferred to go to a theater. In 2014, it was reported that the practice of releasing a film in theaters and via on-demand streaming on the same day (for selected films) and the rise in popularity of the Netflix streaming service has led to concerns in the movie theater industry. Another source of competition is television, which has "...stolen a lot of cinema's best tricks – like good production values and top tier actors – and brought them into people's living rooms". Since the 2010s, one of the increasing sources of competition for movie theaters is the increasing ownership by people of home theater systems which can display high-resolution Blu-ray disks of movies on large, widescreen flat-screen TVs, with 5.1 surround sound and a powerful subwoofer for low-pitched sounds.
Ticket price uniformity
The relatively strong uniformity of movie ticket prices, particularly in the U.S., is a common economics puzzle, because conventional supply and demand theory would suggest higher prices for more popular and more expensive movies, and lower prices for an unpopular "bomb" or for a documentary with less audience appeal. Unlike seemingly similar forms of entertainment such as rock concerts, in which a popular performer's tickets cost much more than an unpopular performer's tickets, the demand for movies is very difficult to predict ahead of time. Indeed, some films with major stars, such as Gigli (which starred the then-supercouple of Ben Affleck and Jennifer Lopez), have turned out to be box-office bombs, while low-budget films with unknown actors have become smash hits (e.g., The Blair Witch Project). The demand for films is usually determined from ticket sale statistics after the movie is already out. Uniform pricing is therefore a strategy to cope with unpredictable demand. Historical and cultural factors are sometimes also cited.
Ticket check
In some movie theater complexes, the theaters are arranged such that tickets are checked at the entrance into the entire plaza, rather than before each theater. At a theater with a sold-out show there is often an additional ticket check, to make sure that everybody with a ticket for that show can find a seat. The lobby may be before or after the ticket check.
Controversies
Advertising: Some moviegoers complain about commercial advertising shorts played before films, arguing that their absence used to be one of the main advantages of going to a movie theater. Other critics such as Roger Ebert have expressed concerns that these advertisements, plus an excessive number of movie trailers, could lead to pressure to restrict the preferred length of the feature films themselves to facilitate playing schedules. So far, the theater companies have typically been highly resistant to these complaints, citing the need for the supplementary income. Some chains like Famous Players and AMC Theatres have compromised with the commercials restricted to being shown before the scheduled start time for the trailers and the feature film. Individual theaters within a chain also sometimes adopt this policy.
Loudness: Another major recent concern is that the dramatic improvements in stereo sound systems and in subwoofer systems have led to cinemas playing the soundtracks of films at unacceptably high volume levels. Usually, the trailers are presented at a very high sound level, presumably to overcome the sounds of a busy crowd. The sound is not adjusted downward for a sparsely occupied theater. Volume is normally adjusted based on the projectionist's judgment of a high or low attendance. The film is usually shown at a lower volume level than the trailers. In response to audience complaints, a manager at a Cinemark theater in California explained that the studios set trailer sound levels, not the theater.
Copyright piracy: In recent years, cinemas have started to show warnings before the movie starts against using cameras and camcorders during the movie (camming). Some patrons record the movie in order to sell "bootleg" copies on the black market. These warnings threaten customers with being removed from the cinema and arrested by the police. This example was shown at cinemas in the United Kingdom:
You are not permitted to use any camera or recording equipment in this cinema. This will be treated as an attempt to breach copyright. Any person doing so can be ejected and such articles may be confiscated by the police. We ask the audience to be vigilant against any such activity and report any matters arousing suspicion to cinema staff. Thank you.
Some theaters (including those with IMAX stadiums) have detectors at the doors to pick up recording smugglers. At particularly anticipated showings, theaters may employ night vision equipment to detect a working camera during a screening. In some jurisdictions this is illegal unless the practice has been announced to the public in advance.
Crowd control: As movie theaters have grown into multiplexes and megaplexes, crowd control has become a major concern. An overcrowded megaplex can be rather unpleasant, and in an emergency can be extremely dangerous (indeed, "shouting fire in a crowded theater" is the standard example of the limits to free speech, because it could cause a deadly panic). Therefore, all major theater chains have implemented crowd control measures. The most well-known measure is the ubiquitous holdout line, which prevents ticket holders for the next showing of that weekend's most popular movie from entering the building until their particular auditorium has been cleared out and cleaned. Since the 1980s, some theater chains (especially AMC Theatres) have developed a policy of co-locating their theaters in shopping centers (as opposed to the old practice of building stand-alone theaters). In some cases, lobbies and corridors cannot hold as many people as the auditoriums, thus making holdout lines necessary. In turn, ticket holders may be enticed to shop or eat while stuck outside in the holdout line. However, given the fact that rent is based on floor area, the practice of having a smaller lobby is somewhat understandable.
Refunds: Most cinema companies issue refunds if there is a technical fault such as a power outage that stops people from seeing a movie. Refunds may be offered during the initial 30 minutes of the screening. The New York Times reported that some audience members walked out of Terrence Malick's film Tree of Life and asked for refunds. At AMC theaters, "...patrons who sat through the entire film and then decided they wanted their money back were out of luck, as AMC's policy is to only offer refunds 30 minutes into a screening. The same goes for Landmark, an independent movie chain... whose policy states, 'If a film is not what is expected… and the feature is viewed less than 30 minutes a refund can be processed for you at the box office.'"
Snack prices: The price of soft drinks and candy at theaters is typically significantly higher than the cost of those items at any other place. Popcorn prices can also be exorbitant. It has been "...estimated that movie theaters make an 85% profit at the concessions stand on overpriced soda, candy, nachos, hot dogs and, of course, popcorn. Movie-theater popcorn has been called one of America's biggest rip-offs, with a retail price of nine times what it costs to make."
Cinema and movie theater chains
In Africa, Ster-Kinekor has the largest market share in South Africa. Nu Metro Cinemas is another cinema chain in South Africa.
In North America, the National Association of Theatre Owners (NATO) is the largest exhibition trade organization in the world. According to their figures, the top four chains represent almost half of the theater screens in North America. In Canada, Cineplex Entertainment is by far the largest player with 161 locations and 1,635 screens.
In the United States, the studios once controlled many theaters, but after the appearance of Mr. Smith Goes to Washington, Congress passed the Neely Anti-Block Booking Act, which eventually broke the link between the studios and the theaters. Now, the top three chains in the U.S. are Regal Entertainment Group, AMC Entertainment Inc and Cinemark Theatres. In 1995, Carmike was the largest chain in the United States- now, the major chains include AMC Entertainment Inc – 5,206 screens in 346 theaters, Cinemark Theatres – 4,457 screens in 334 theaters, Landmark Theatres – 220 screens in 54 theaters, Marcus Theatres – 681 screens in 53 theaters. National Amusements – 409 screens in 32 theaters and Regal Entertainment Group – 7,334 screens in 588 cinemas. In 2015 the United States had a total of 40,547 screens. In Mexico, the major chains are Cinepolis and Cinemex.
In South America, Argentine chains include Hoyts, Village Cinemas, Cinemark and Showcase Cinemas. Brazilian chains include Cinemark and Moviecom. Chilean chains include Hoyts and Cinemark. Colombian, Costa Rican, Panamanian and Peruvian chains include Cinemark and Cinépolis.
In Asia, Wanda Cinemas is the largest exhibitor in China, with 2,700 screens in 311 theaters and with 18% of the screens in the country; another major Chinese chain is UA Cinemas. China had a total of 31,627 screens in 2015 and is expected to have almost 40,000 in 2016. Hong Kong has AMC Theatres. South Korea's CJ CGV also has branches in China, Indonesia, Myanmar, Turkey, Vietnam, and the United States. In India, PVR Cinemas is a leading cinema operating a chain of 500 screens and CineMAX and INOX are both multiplex chains. These theatres practice safety guidelines in each cinema halls. Indonesia has the 21 Cineplex and Cinemaxx (As od 2019, renamed as Cinépolis) chain. A major Israel theater is Cinema City International. Japanese chains include Toho and Shochiku.
Europe is served by AMC, Cineworld, Vue Cinema and Odeon.
In Oceania (particularly Australia), large chains include Event Cinemas, Village Cinemas, Hoyts Cinemas and Palace Cinemas.
See also
Cinema etiquette
Film poster
Film screening
Home cinema
Inflatable movie screen
List of drive-in theaters
List of movie theaters
Mini theater
Movie palace
Multiplex (movie theater)
Nickelodeon (movie theater)
Saturday morning pictures
References
External links
Movie theaters early users of air conditioning – Pantagraph (Bloomington, Illinois newspaper)
Audiovisual introductions in 1895
Audiovisual introductions in 1896
Audiovisual introductions in 1986
Audiovisual introductions in 2016
Auditoriums
Promotion and marketing communications
Theatres |
null | null | Didymus | eng_Latn | didymus (Greek for "twin") may refer to:
People
Didymus
Arius Didymus (fl. 1st century BC), Stoic philosopher
Thomas the Apostle (died 72 AD), also known as Didymus
Didymus Chalcenterus (63 BC–10 AD), Hellenistic scholar and grammarian
Didymus the Blind (313–398), ecclesiastical writer of Alexandria
Didymus the Musician, music theorist in Alexandria of the 1st century
St Didymus (died 304), 4th century companion of StTheodora
Didymus, a member of the House of Theodosius
Didymus Mutasa (born 1935), Minister of National Security in Zimbabwe
Didymos
Didymos (music theorist), ancient Greek music theorist
Didymos I (1921–2014), Catholicos of the East and Malankara Metropolitan, the Primate of the Malankara Church of India from 2005 to 2010.
Fictional characters
Sir Didymus, a fictional character from the movie Labyrinth
Animals
Didymus (beetle), a genus of weevils found in New Zealand, the Kermadec and Norfolk Islands
Leptodactylus didymus, a species of frog in Bolivia and Peru
See also
Didymos (disambiguation)
Didimus, a genus of beetle
65803 Didymos, a near-Earth asteroid |
null | null | North Platte, Nebraska | eng_Latn | North Platte is a city in and the county seat of Lincoln County, Nebraska, United States. It is located in the west-central part of the state, along Interstate 80, at the confluence of the North and South Platte Rivers forming the Platte River. The population was 23,390 at the 2020 census.
North Platte is a railroad town; Union Pacific Railroad's large Bailey Yard is located within the city. Today, North Platte is served only by freight trains, but during World War II the city was known for the North Platte Canteen, a volunteer organization serving food to millions of traveling soldiers.
North Platte is the principal city of the North Platte Micropolitan Statistical Area, which includes Lincoln, Logan, and McPherson counties.
History
North Platte was established in 1866 when the Union Pacific Railroad was extended to that point. It derives its name from the North Platte River.
North Platte was the western terminus of the Union Pacific Railway from the summer of 1867 until the next section to Laramie, Wyoming, was opened the following summer. Even though Congress had authorized the building of the Transcontinental Railroad in 1862, it had been extended only as far as Nebraska City by the start of the summer of 1867. The section from Nebraska City to North Platte was completed in less than six weeks.
In the 1880s, Buffalo Bill Cody established his ranch, known as Scout's Rest Ranch, just north of town. It is now a National Historic Landmark.
On July 13, 1929, a black man shot and killed a white police officer. The black man reportedly took his own life, being trapped by a mob. This led to the formation of white mobs combing the city, and ordering black residents to leave North Platte. Fearing mob violence, most of North Platte's black residents fled.
The North Platte Canteen was one of the largest volunteer efforts of World War II, originating in 1941. Tens of thousands of volunteers from North Platte and surrounding towns met the troop trains passing through North Platte, offering coffee, sandwiches, dessert, and hospitality to nearly seven million servicemen.
Geography
According to the United States Census Bureau, the city has a total area of , of which is land and is water.
Climate
North Platte experiences a dry continental climate similar to that of the Nebraska High Plains, classified as hot-summer humid continental (Köppen Dfa/Dwa), and, with an annual average precipitation of , barely avoids semi-arid classification; it is part of USDA Hardiness zone 5a. The normal monthly mean temperature ranges from in January to in July. On an average year, there are 3.8 afternoons that reach or higher, 39 afternoons that reach or higher, 31.4 afternoons that do not climb above freezing, and 12.2 mornings with a low of or below. The average window for freezing temperatures is September 30 thru May 13, allowing a growing season of 139 days. Extreme temperatures officially range from on January 15, 1888 and February 12, 1899, up to on July 11, 1954; the record cold daily maximum is on January 14, 1888, while, conversely, the record warm daily minimum is on July 25, 1940.
Precipitation is greatest in May and June and has ranged from in 1931 to in 1951. Snowfall averages per season, and has historically ranged from in 1903–04 to in 1979–80; the average window for measurable (≥) snowfall is November 1 thru April 12, with May snow being rare.
Demographics
2010 census
As of the census of 2010, there were 24,733 people, 10,560 households, and 6,290 families residing in the city. The population density was . There were 11,450 housing units at an average density of . The racial makeup of the city was 93.1% White, 1.0% African American, 0.7% Native American, 0.7% Asian, 2.8% from other races, and 1.7% from two or more races. Hispanic or Latino of any race were 8.8% of the population.
There were 10,560 households, of which 30.4% had children under the age of 18 living with them, 44.5% were married couples living together, 10.7% had a female householder with no husband present, 4.3% had a male householder with no wife present, and 40.4% were non-families. 34.8% of all households were made up of individuals, and 13.9% had someone living alone who was 65 years of age or older. The average household size was 2.29 and the average family size was 2.95.
The median age in the city was 37.1 years. 24.9% of residents were under the age of 18; 9% were between the ages of 18 and 24; 25% were from 25 to 44; 25.6% were from 45 to 64; and 15.5% were 65 years of age or older. The gender makeup of the city was 48.8% male and 51.2% female.
2000 census
As of the census of 2000, there were 23,878 people, 9,944 households, and 6,224 families residing in the city. The population density was 2,281.5 people per square mile (880.5/km2). There were 10,718 housing units at an average density of 1,024.1 per square mile (395.2/km2). The racial makeup of the city was 93.47% White, 0.71% African American, 0.64% Native American, 0.39% Asian, 0.03% Pacific Islander, 3.30% from other races, and 1.45% from two or more races. Hispanic or Latino of any race were 6.68% of the population.
There were 9,944 households, out of which 31.0% had children under the age of 18 living with them, 49.8% were married couples living together, 9.6% had a female householder with no husband present, and 37.4% were non-families. 31.9% of all households were made up of individuals, and 13.0% had someone living alone who was 65 years of age or older. The average household size was 2.34 and the average family size was 2.97.
In the city, the population was spread out, with 26.0% under the age of 18, 9.5% from 18 to 24, 26.8% from 25 to 44, 21.9% from 45 to 64, and 15.8% who were 65 years of age or older. The median age was 36 years. For every 100 females, there were 94.5 males. For every 100 females age 18 and over, there were 90.0 males.
As of 2000 the median income for a household in the city was $34,181, and the median income for a family was $42,753. Males had a median income of $36,445 versus $20,157 for females. The per capita income for the city was $18,306. About 7.8% of families and 10.5% of the population were below the poverty line, including 13.2% of those under age 18 and 9.8% of those age 65 or over.
Arts and culture
Lincoln County Historical Museum contains a display detailing the history of the North Platte Canteen. It also contains a Prairie Village with local landmark homes and other buildings, including a Pony Express station and pioneer church among many others.
Buffalo Bill Ranch State Historical Park is located near North Platte, a Nebraska living history park about "Buffalo Bill" Cody. The park includes his actual house known as Scout's Rest Ranch. The park is two miles west of U.S. Highway 83 along U.S. Highway 30.
Every June, North Platte hosts the annual "Nebraskaland Days". The event includes parades, art shows, rodeos, concerts, and food events. It draws over 100,000 attendees every year.
North Platte is host to the annual Miss Nebraska pageant, an official preliminary for the Miss America Organization.
Infrastructure
Transportation
North Platte is home to the world's largest rail yard, Bailey Yard. The Golden Spike Tower and Visitor Center is an eight-story building which overlooks the expansive classification yard and engine facilities. The tower and visitor center are open to the public year-round. Passenger train service was discontinued in 1971.
North Platte is home to North Platte Regional Airport. United Express serves the airport with twice-daily service to Denver International Airport. There is also a door-to-door bus system available for residents of the town.
Media
Notable people
Chris Başak, baseball infielder was born in North Platte. He played in the New York Mets, Minnesota Twins and New York Yankees organizations. He made five appearances in the major leagues in 2007 with the Yankees, going 0-for-1.
Howard Baskerville, American teacher in the Presbyterian mission school in Tabriz, Iran; born in North Platte. He was killed defending the Constitutional Revolution of Iran and known as "American Martyr of the Constitutional Revolution."
Buffalo Bill Cody, iconic Wild West showman, owned a ranch in North Platte.
Nathan Enderle, football quarterback, selected by the Chicago Bears in the fifth round of the 2011 NFL Draft; born in North Platte and attended North Platte High School.
Paul Faulkner (1913–1997), artist; born in North Platte.
Chuck Hagel, Nebraska U.S. Senator and U.S. Secretary of Defense; born in North Platte.
Bill Hayes, baseball player, lived in North Platte and attended Saint Patrick's High School.
Henry Hill (1943–2012), New York City mobster; worked as a cook in North Platte.
John Howell, former American football safety in the National Football League, played for the Tampa Bay Buccaneers and Seattle Seahawks; born in North Platte.
Glenn Miller, popular big band leader; lived in North Platte for one year during his early childhood.
Keith Neville, 18th governor of Nebraska; born in North Platte.
Joe Ragland (born 1989), American-Liberian basketball player for Hapoel Holon of the Israeli Basketball Premier League
Red Cloud, Sioux warrior; born near North Platte in 1822.
Dr. Don Rose, San Francisco Bay Area disc jockey; born and raised in North Platte and frequently referred to the city on his morning show.
Ryan Schultz, professional mixed martial artist, Lightweight Champion of defunct IFL; raised in North Platte.
Zane Smith, professional baseball player, was a two-sport athlete for North Platte 1978-79, before spending 13 years in major league baseball.
Danny Woodhead, retired NFL running back; born in North Platte and attended North Platte High School.
See also
National Register of Historic Places in Lincoln County, Nebraska
Explanatory notes
References
Further reading
External links
City of North Platte
North Platte Area Chamber of Commerce
North Platte Lincoln County Convention and Visitors Bureau
Cities in Lincoln County, Nebraska
Cities in Nebraska
North Platte Micropolitan Statistical Area
County seats in Nebraska
1868 establishments in Nebraska |
null | null | United Nations Development Programme | eng_Latn | The United Nations Development Programme (UNDP) is a United Nations organization tasked with helping countries eliminate poverty and achieve sustainable economic growth and human development. Headquartered in New York City, it is the largest UN development aid agency, with offices in 170 countries.
The UNDP emphasizes developing local capacity towards long-term self-sufficiency and prosperity. It administers projects to attract investment, technical training, and technological development, and provides experts to help build legal and political institutions and expand the private sector. Assistance is provided at the request of national governments, and UNDP personnel frequently coordinate with local officials, nongovernmental organizations, and other international agencies.
The UNDP operates in 125 counties and is funded entirely by voluntary contributions from UN member states. In addition to meeting national development objectives, its activities increasingly focus on fulfilling the Sustainable Development Goals (SDGs), such as combating HIV/AIDS and ensuring environmentally sustainable development. The UNDP is governed by a 36-member executive board overseen by an administrator, who is third-highest ranking UN official after the Secretary-General and Deputy Secretary-General.
Founding
The UNDP was founded on 22 November 1965 with the merging of the Expanded Programme of Technical Assistance (EPTA) and the Special Fund. The rationale was to "avoid duplication of [their] activities". The EPTA was set up in 1949 to help the economic and political aspects of underdeveloped countries while the Special Fund was to enlarge the scope of UN technical assistance. The Special Fund arose from the idea of a Special United Nations Fund for Economic Development (SUNFED) (which was initially called the United Nations Fund for Economic Development (UNFED).
Countries such as the Nordic countries were proponents of such a United Nations (UN) controlled fund. However, the fund was opposed by developed countries, especially by the United States who was wary of the Third World dominating such a funding and preferred it to be under the auspices of the World Bank. The concept of SUNFED was dropped to form the Special Fund. This Special Fund was a compromise over the SUNFED concept, it did not provide investment capital, but only helped to bring pre-conditions for private investment.
With the US proposing and creating the International Development Association within the World Bank's umbrella, the EPTA and the Special Fund appeared to be conducting similar work. In 1962, the United Nations Economic and Social Council asked the Secretary General to consider the merits and disadvantages of merging UN technical assistance programs and in 1966, the EPTA and the Special Fund merged to form the UNDP.
Budget
In 2019, UNDP's entire budget was approximately US$6 billion.
Funding information table
The following table lists the top 15 DAC 5 Digit Sectors to which UNDP has committed funding, as recorded in its International Aid Transparency Initiative (IATI) publications. The UNDP claims on the IATI Registry website that the data covers 100% of development flows.
UNDP topped the Aid Transparency Index published by Publish What You Fund in 2015 and 2016. "The United Nations Development Programme (UNDP) tops the Index for the second time with an excellent score of 93.3%, the only organisation to score above 90%".
UNDP links and coordinates global and national efforts to achieve the goals and national development priorities laid out by host countries. UNDP focuses primarily on five developmental challenges:
Democratic governance
UNDP supports national democratic transitions by providing policy advice and technical support, improving institutional and individual capacity within countries, educating populations about and advocating for democratic reforms, promoting negotiation and dialogue, and sharing successful experiences from other countries and locations. UNDP also supports existing democratic institutions by increasing dialogue, enhancing national debate, and facilitating consensus on national governance programmes.
Poverty reduction
UNDP helps countries develop strategies to combat poverty by expanding access to economic opportunities and resources, linking poverty programmes with countries’ larger goals and policies, and ensuring a greater voice for the poor. It also works at the macro level to reform trade, encourage debt relief and foreign investment, and ensure the poorest of the poor benefit from globalisation. On the ground, UNDP sponsors developmental pilot projects, promotes the role of women in development, and coordinates efforts between governments, NGOs, and outside donors. In this way, UNDP works with local leaders and governments to provide opportunities for impoverished people to create businesses and improve their economic condition.
The UNDP united nation development programme International Policy Centre for Inclusive Growth (IPC-IG) in Brasília, Brazil, expands the capacities of developing countries to design, implement and evaluate socially inclusive development projects. IPC-IG is a global forum for South-South policy dialogue and learning, having worked with more than 7,000 officials from more than 50 countries.
A 2013 evaluation of the UNDP's poverty reduction efforts states that the UNDP has effectively supported national efforts to reduce poverty, by helping governments make policy changes that benefit the poor. Nevertheless, the same evaluation also states there is a strong need for better measurement and monitoring of the impacts of the UNDP's work. The UNDP's Strategic Plan from 2014 to 2017 incorporates the recommendations of this poverty evaluation.
Crisis prevention and recovery
UNDP works to reduce the risk of armed conflicts or disasters, and promote early recovery after crisis have occurred. UNDP works through its country offices to support local government in needs assessment, capacity development, coordinated planning, and policy and standard setting.
Examples of UNDP risk reduction programmes include efforts to control small arms proliferation, strategies to reduce the impact of natural disasters, and programmes to encourage use of diplomacy and prevent violence. Recovery programmes include disarmament, demobilization and reintegration of ex-combatants, demining efforts, programmes to reintegrate displaced persons, restoration of basic services, and transitional justice systems for countries recovering from warfare.
Following the suspension of most foreign aid to Afghanistan due to its takeover by the Taliban, the UNDP took responsibility for funding most essential health services in the country, including the salaries of over 25,00 health care professionals. This was observed as being outside the organization's usual development activities, and was facilitated by special licensing by the United States government.
Environment and energy
As the poor are disproportionately affected by environmental degradation and lack of access to clean, affordable water, sanitation, and energy services, UNDP seeks to address environmental issues in order to improve developing countries’ abilities to develop sustainably, increase human development and reduce poverty. UNDP works with countries to strengthen their capacity to address global environmental issues by providing innovative policy advice and linking partners through environmentally sensitive development projects that help poor people build sustainable livelihoods.
UNDP's environmental strategy focuses on effective water governance including access to water supply and sanitation, access to sustainable energy services, Sustainable land management to combat desertification and land degradation, conservation and sustainable use of biodiversity, and policies to control emissions of harmful pollutants and ozone-depleting substances. UNDP's Equator Initiative office biennially offers the Equator Prize to recognize outstanding indigenous community efforts to reduce poverty through the conservation and sustainable use of biodiversity, and thus making local contributions to achieving the Sustainable Development Goals (SDGs).
Between 1996 and 1998, the UNDP sponsored the deployment of 45 Multifunction Platforms (MFP) in rural Mali. These installations, driven by a diesel engine, power devices such as pumps, grain mills and appliances. By 2004, the number of MFPs in Mali reached 500.
In 2012 the Biodiversity Finance Initiative (BIOFIN) was established. BIOFIN brings 30 countries together to develop and implement evidence-based finance plans to safeguards biodiversity. BIOFIN has developed an innovative and adaptable methodology to guide countries to analyze the policy and institutional context for biodiversity finance; measure the current biodiversity expenditures; assess future financial needs, and identify the most suitable finance solutions to achieve national biodiversity targets.
HIV/AIDS
UNDP works to help countries prevent further spreading of and reduce the impact of HIV/AIDS, convening The Global Commission on HIV and the Law which reported in 2012.
Hub for Innovative Partnerships
Major programmes underway are:
ART Global Initiative
World Alliance of Cities Against Poverty
Territorial Approach to Climate Change
Africa–Kazakhstan Partnership for the SDGs
Human Development Report
Since 1991, the UNDP has annually published the Human Development Report, which includes topics on Human Development and the annual Human Development Index.
The Gender Inequality Index is one such topic discussed in the Human Development Report.
Evaluation
The UNDP spends about 0.2% of its budget on internal evaluation of the effectiveness of its programmes. The UNDP's Evaluation Office is a member of the UN Evaluation Group (UNEG) which brings together all the units responsible for evaluation in the UN system. Currently the UNEG has 43 members and 3 observers.
Global Policy Centers
The UNDP runs six GPCs, including the Seoul GPC on partnerships, and the Global Center for Public Service Excellence that issues the "Raffles Review" email newsletter on developments in public administration research.
UN coordination role
UNDP plays a significant coordination role for the UN's activities in the field of development. This is mainly executed through its leadership of the UN Development Group and through the Resident Coordinator System.
United Nations Development Group
The United Nations Development Group (UNDG) was created by the Secretary-General in 1997, to improve the effectiveness of UN development at the country level. The UNDG brings together the operational agencies working on development. The Group is chaired by the Administrator of UNDP. UNDP also provides the Secretariat to the Group.
The UNDG develops policies and procedures that allow member agencies to work together and analyze country issues, plan support strategies, implement support programs, monitor results and advocate for change. These initiatives increase UN impact in helping countries achieve the Millennium Development Goals (MDGs), including poverty reduction.
Thirty-two UN agencies are members of the UNDG. The executive committee consists of the four "founding members": UNICEF, UNFPA, WFP and UNDP. The Office of the High Commissioner for Human Rights is an ex-officio member of the executive committee.
Resident coordinator system
The Resident Coordinator (RC) system coordinates all organizations of the United Nations system dealing with operational activities for development in the field. The RC system aims to bring together the different UN agencies to improve the efficiency and effectiveness of operational activities at the country level. Resident Coordinators lead UN country teams in more than 130 countries and are the designated representatives of the Secretary-General for development operations. Working closely with national governments, Resident Coordinators and country teams advocate the interests and mandates of the UN drawing on the support and guidance of the entire UN family. It is now coordinated by the UNDG.
Innovation Facility
The UNDP established the Innovation Facility in 2014, with support from the Government of Denmark, as a dedicated funding mechanism to nurture promising development interventions.
The Innovation Facility offers technical assistance and seed funding to collaborators across 170 countries and territories to explore new approaches to complex development challenges. Since its inception, the Innovation Facility has fostered innovation labs across all 5 regions to better deliver and monitor SDGs. In 2015, the Innovation Facility invested in 62 initiatives across 45 countries to achieve 16 SDGs.
Controversies
NSA surveillance
Documents of Edward Snowden showed in December 2013 that British and American intelligence agencies surveillance targets with the United States' National Security Agency (NSA) included organisations such as the United Nations Development Programme, the UN's children's charity UNICEF and Médecins Sans Frontières and the Economic Community of West African States (ECOWAS).
Allegations of UNDP resources used by Hamas
In August 2016, Israel's Shin Bet security agency publicized the arrest of Wahid Abdallah al Bursh, a Palestinian engineer employed by the UNDP, stating he had confessed to being recruited in 2014 to help Hamas, the dominant Islamist group in Gaza. Among "various assignments" Bursh performed on behalf of Hamas was "using UNDP resources" to build a maritime jetty for its fighters; no further details were provided on this claim. Shin Bet also claimed Bursh had persuaded his UNDP superiors to prioritize neighborhoods with Hamas operatives when earmarking reconstruction funds for Gaza, which was devastated by the 2014 war with Israel.
Alleged financial irregularities
The UNDP had been criticised by members of its staff and the Bush administration of the United States for irregularities in its finances in North Korea. Artjon Shkurtaj claimed that he had found counterfeit US dollars in the programmes' safe despite staff being paid in euros. The UNDP denied keeping improper accounts and any other wrongdoing.
Disarmament and controversy
In mid-2006, as first reported by Inner City Press and then by New Vision, UNDP halted its disarmament programmes in the Karamoja region of Uganda in response to human rights abuses in the parallel forcible disarmament programmes carried out by the Uganda People's Defence Force.
Administrator
The UNDP Administrator has the rank of an Under-Secretary-General of the United Nations. While the Administrator is often referred to as the third highest-ranking official in the UN (after the UN Secretary General and the UN Deputy Secretary General), this has never been formally codified.
In addition to his/her responsibilities as head of UNDP, the Administrator is also the vice-chair of the UN Sustainable Development Group.
The position of Administrator is appointed by the Secretary-General of the UN and confirmed by the General Assembly for a term of four years.
Achim Steiner is the current Administrator. The five countries on the UNDP board have some influence over selection of the administrator.
The first administrator of the UNDP was Paul G. Hoffman, former head of the Economic Cooperation Administration which administered the Marshall Plan.
Other holders of the position have included: Bradford Morse, former Republican congressman from Massachusetts; William Draper, venture capitalist and friend of George H.W. Bush who saw one of the UN system's major achievements, the Human Development Report, introduced during his tenure; Mark Malloch Brown, who was previously Vice President of External Affairs at the World Bank and subsequently became UN Deputy Secretary General. Kemal Derviş, a former finance minister of Turkey and senior World Bank official, was the previous UNDP Administrator. Derviş started his four-year term on 15 August 2005.
Associate Administrator
During meetings of the UN Sustainable Development Group, which are chaired by the Administrator, UNDP is represented by the Associate Administrator. The position is currently held by Tegegnework Gettu, appointed on 1 December 2015. He is also currently service as Director a.i., for the Africa Region.
Assistant administrators
Assistant Administrators of the UNDP, Assistant United Nations Secretaries General and Directors of the Regional Bureaus are
Usha Rao-Monari (India), USG and Associate Administrator;
Khalida Bouzar- Algeria- UN Assistant Secretary General, UNDP Assistant Administrator and Director of the UNDP
Regional Bureau for Arab States
Abdoulaye Mar Dieye (Senegal), for Bureau for Programme and Policy Support;
Haoliang Xu (China), for Asia & Pacific;
Mirjana Spoljaric Egger (Switzerland), for Europe & CIS and
Luis Felipe López Calva (Mexico), for Latin America and the Caribbean.
See also
United Nations Sustainable Development Group
UNDP Goodwill Ambassador
Human development
Development assistance
Economic development
ECOSOC
International Covenant on Economic, Social and Cultural Rights
International development
International Development Association
List of UNDP country codes
List of HIV-positive people
U.S. Committee for the United Nations Development Program
United Nations Sustainable Development Group
United Nations Millennium Campaign
United Nations Volunteers
Equator Prize
Match Against Poverty
United Nations Volunteers
World Population Day
Notes
References
External links
Official UNDP web site
Jackson, R. G. A., A Study of the Capacity of the United Nations Development System. 2 vols, Geneva: United Nations, 1969.
Mitcham, Chad J. ‘Australia and Development Cooperation at the United Nations: Towards Poverty Reduction.’ In Australia and the United Nations, edited by James Cotton and David Lee, 191–221. Canberra: Department of Foreign Affairs and Trade and Sydney: Longueville Books, 2013.
Mitcham, Chad J., Jackson, Sir Robert Gillman (1911-1991), Australian Dictionary of Biography, National Centre of Biography, Australian National University, http://adb.anu.edu.au/biography/jackson-sir-robert-gillman-20715/text31511, published online 2016, accessed online 5 September 2017.
International development organizations
United Nations General Assembly subsidiary organs
Organizations based in New York City
United Nations Development Group
1965 establishments in the United States
1965 establishments in New York (state)
Economic development |
null | null | Theatre of the Absurd | eng_Latn | The Theatre of the Absurd ( ) is a post–World War II designation for particular plays of absurdist fiction written by a number of primarily European playwrights in the late 1950s. It is also a term for the style of theatre the plays represent. The plays focus largely on ideas of existentialism and express what happens when human existence lacks meaning or purpose and communication breaks down. The structure of the plays is typically a round shape, with the finishing point the same as the starting point. Logical construction and argument give way to irrational and illogical speech and to the ultimate conclusion—silence.
Etymology
Critic Martin Esslin coined the term in his 1960 essay "The Theatre of the Absurd", which begins by focusing on the playwrights Samuel Beckett, Arthur Adamov, and Eugène Ionesco. Esslin says that their plays have a common denominator — the "absurd", a word that Esslin defines with a quotation from Ionesco: "absurd is that which has not purpose, or goal, or objective." The French philosopher Albert Camus, in his 1942 essay "Myth of Sisyphus", describes the human situation as meaningless and absurd.
The Absurd in these plays takes the form of man's reaction to a world apparently without meaning, or man as a puppet controlled or menaced by invisible outside forces. This style of writing was first popularized by the Eugène Ionesco play The Bald Soprano (1950). Although the term is applied to a wide range of plays, some characteristics coincide in many of the plays: broad comedy, often similar to vaudeville, mixed with horrific or tragic images; characters caught in hopeless situations forced to do repetitive or meaningless actions; dialogue full of clichés, wordplay, and nonsense; plots that are cyclical or absurdly expansive; either a parody or dismissal of realism and the concept of the "well-made play".
In his introduction to the book Absurd Drama (1965), Esslin wrote:The Theatre of the Absurd attacks the comfortable certainties of religious or political orthodoxy. It aims to shock its audience out of complacency, to bring it face to face with the harsh facts of the human situation as these writers see it. But the challenge behind this message is anything but one of despair. It is a challenge to accept the human condition as it is, in all its mystery and absurdity, and to bear it with dignity, nobly, responsibly; precisely there are no easy solutions to the mysteries of existence, because ultimately man is alone in a meaningless world. The shedding of easy solutions, of comforting illusions, may be painful, but it leaves behind it a sense of freedom and relief. And that is why, in the last resort, the Theatre of the Absurd does not provoke tears of despair but the laughter of liberation.
Origin
In the first edition of The Theatre of the Absurd, Esslin quotes the French philosopher Albert Camus's essay "Myth of Sisyphus", as it uses the word "absurdity" to describe the human situation: "In a universe that is suddenly deprived of illusions and of light, man feels a stranger. … This divorce between man and his life, the actor and his setting, truly constitutes the feeling of Absurdity."
Esslin presents the four defining playwrights of the movement as Samuel Beckett, Arthur Adamov, Eugène Ionesco, and Jean Genet, and in subsequent editions he added a fifth playwright, Harold Pinter. Other writers associated with this group by Esslin and other critics include Tom Stoppard, Friedrich Dürrenmatt, Fernando Arrabal, Edward Albee, Boris Vian, and Jean Tardieu.
Precursors
Elizabethan – tragicomedy
The mode of most "absurdist" plays is tragicomedy. As Nell says in Endgame, "Nothing is funnier than unhappiness … it's the most comical thing in the world". Esslin cites William Shakespeare as an influence on this aspect of the "Absurd drama". Shakespeare's influence is acknowledged directly in the titles of Ionesco's Macbett and Stoppard's Rosencrantz and Guildenstern Are Dead. Friedrich Dürrenmatt says in his essay "Problems of the Theatre", "Comedy alone is suitable for us … But the tragic is still possible even if pure tragedy is not. We can achieve the tragic out of comedy. We can bring it forth as a frightening moment, as an abyss that opens suddenly; indeed, many of Shakespeare's tragedies are already really comedies out of which the tragic arises."
Though layered with a significant amount of tragedy, Theatre of the Absurd echoes other great forms of comedic performance, according to Esslin, from Commedia dell'arte to vaudeville. Similarly, Esslin cites early film comedians and music hall artists such as Charlie Chaplin, the Keystone Cops and Buster Keaton as direct influences. (Keaton even starred in Beckett's Film in 1965.)
Formal experimentation
As an experimental form of theatre, many Theatre of the Absurd playwrights employ techniques borrowed from earlier innovators. Writers and techniques frequently mentioned in relation to the Theatre of the Absurd include the 19th-century nonsense poets, such as Lewis Carroll or Edward Lear; Polish playwright Stanisław Ignacy Witkiewicz; the Russians Daniil Kharms, Nikolai Erdman, and others; Bertolt Brecht's distancing techniques in his "Epic theatre"; and the "dream plays" of August Strindberg.
One commonly cited precursor is Luigi Pirandello, especially Six Characters in Search of an Author. Pirandello was a highly regarded theatrical experimentalist who wanted to bring down the fourth wall presupposed by the realism of playwrights such as Henrik Ibsen. According to W. B. Worthen, Six Characters and other Pirandello plays use "Metatheatre—roleplaying, plays-within-plays, and a flexible sense of the limits of stage and illusion—to examine a highly-theatricalized vision of identity".
Another influential playwright was Guillaume Apollinaire whose The Breasts of Tiresias was the first work to be called "surreal".
Pataphysics, surrealism, and Dadaism
A precursor is Alfred Jarry whose Ubu plays scandalized Paris in the 1890s. Likewise, the concept of 'pataphysics—"the science of imaginary solutions"—first presented in Jarry's Gestes et opinions du docteur Faustroll, pataphysicien (Exploits and Opinions of Dr. Faustroll, pataphysician) was inspirational to many later Absurdists, some of whom joined the Collège de 'pataphysique, founded in honor of Jarry in 1948 (Ionesco, Arrabal, and Vian were given the title Transcendent Satrape of the Collège de 'pataphysique). The Theatre Alfred Jarry, founded by Antonin Artaud and Roger Vitrac, housed several Absurdist plays, including ones by Ionesco and Adamov.
Artaud's "The Theatre of Cruelty" (presented in The Theatre and Its Double) was a particularly important philosophical treatise. Artaud claimed theatre's reliance on literature was inadequate and that the true power of theatre was in its visceral impact. Artaud was a Surrealist, and many other members of the Surrealist group were significant influences on the Absurdists.
Absurdism is also frequently compared to Surrealism's predecessor, Dadaism (for example, the Dadaist plays by Tristan Tzara performed at the Cabaret Voltaire in Zürich). Many of the Absurdists had direct connections with the Dadaists and Surrealists. Ionesco, Adamov, and Arrabal for example, were friends with Surrealists still living in Paris at the time including Paul Eluard and André Breton, the founder of Surrealism, and Beckett translated many Surrealist poems by Breton and others from French into English.
Relationship with existentialism
Many of the Absurdists were contemporaries with Jean-Paul Sartre, the philosophical spokesman for existentialism in Paris, but few Absurdists actually committed to Sartre's own existentialist philosophy, as expressed in Being and Nothingness, and many of the Absurdists had a complicated relationship with him. Sartre praised Genet's plays, stating that for Genet, "Good is only an illusion. Evil is a Nothingness which arises upon the ruins of Good".
Ionesco, however, hated Sartre bitterly. Ionesco accused Sartre of supporting Communism but ignoring the atrocities committed by Communists; he wrote Rhinoceros as a criticism of blind conformity, whether it be to Nazism or Communism; at the end of the play, one man remains on Earth resisting transformation into a rhinoceros. Sartre criticized Rhinoceros by questioning: "Why is there one man who resists? At least we could learn why, but no, we learn not even that. He resists because he is there". Sartre's criticism highlights a primary difference between the Theatre of the Absurd and existentialism: the Theatre of the Absurd shows the failure of man without recommending a solution. In a 1966 interview, Claude Bonnefoy, comparing the Absurdists to Sartre and Camus, said to Ionesco, "It seems to me that Beckett, Adamov and yourself started out less from philosophical reflections or a return to classical sources, than from first-hand experience and a desire to find a new theatrical expression that would enable you to render this experience in all its acuteness and also its immediacy. If Sartre and Camus thought out these themes, you expressed them in a far more vital contemporary fashion". Ionesco replied, "I have the feeling that these writers – who are serious and important – were talking about absurdity and death, but that they never really lived these themes, that they did not feel them within themselves in an almost irrational, visceral way, that all this was not deeply inscribed in their language. With them it was still rhetoric, eloquence. With Adamov and Beckett it really is a very naked reality that is conveyed through the apparent dislocation of language".
In comparison to Sartre's concepts of the function of literature, Samuel Beckett's primary focus was on the failure of man to overcome "absurdity" - or the repetition of life even though the end result will be the same no matter what and everything is essentially pointless - as James Knowlson says in Damned to Fame, Beckett's work focuses, "on poverty, failure, exile and loss — as he put it, on man as a 'non-knower' and as a 'non-can-er' ." Beckett's own relationship with Sartre was complicated by a mistake made in the publication of one of his stories in Sartre's journal Les Temps Modernes. Beckett said, though he liked Nausea, he generally found the writing style of Sartre and Heidegger to be "too philosophical" and he considered himself "not a philosopher".
History
The "Absurd" or "New Theater" movement was originally a Paris-based (and a Rive Gauche) avant-garde phenomenon tied to extremely small theaters in the Quartier Latin. Some of the Absurdists, such as Jean Genet, Jean Tardieu, and Boris Vian., were born in France. Many other Absurdists were born elsewhere but lived in France, writing often in French: Samuel Beckett from Ireland; Eugène Ionesco from Romania; Arthur Adamov from Russia; Alejandro Jodorowsky from Chile and Fernando Arrabal from Spain. As the influence of the Absurdists grew, the style spread to other countries—with playwrights either directly influenced by Absurdists in Paris or playwrights labelled Absurdist by critics. In England, some of those whom Esslin considered practitioners of the Theatre of the Absurd include Harold Pinter, Tom Stoppard, N. F. Simpson, James Saunders, and David Campton; in the United States, Edward Albee, Sam Shepard, Jack Gelber, and John Guare; in Poland, Tadeusz Różewicz; Sławomir Mrożek, and Tadeusz Kantor; in Italy, Dino Buzzati; and in Germany, Peter Weiss, Wolfgang Hildesheimer, and Günter Grass. In India, both Mohit Chattopadhyay and Mahesh Elkunchwar have also been labeled Absurdists. Other international Absurdist playwrights include Tawfiq el-Hakim from Egypt; Hanoch Levin from Israel; Miguel Mihura from Spain; José de Almada Negreiros from Portugal; Mikhail Volokhov from Russia; Yordan Radichkov from Bulgaria; and playwright and former Czech President Václav Havel.
Major productions
Jean Genet's The Maids (Les Bonnes) premiered in 1947.
Eugène Ionesco's The Bald Soprano (La Cantatrice Chauve) was first performed on May 11, 1950, at the Théâtre des Noctambules. Ionesco followed this with The Lesson (La Leçon) in 1951 and The Chairs (Les Chaises) in 1952.
Samuel Beckett's Waiting for Godot was first performed on 5 January 1953 at the Théâtre de Babylone in Paris.
In 1957, Genet's The Balcony (Le Balcon) was produced in London at the Arts Theatre.
That May, Harold Pinter's The Room was presented at The Drama Studio at the University of Bristol. Pinter's The Birthday Party premiered in the West End in 1958.
Edward Albee's The Zoo Story premiered in West Berlin at the Schiller Theater Werkstatt in 1959.
On October 28, 1959, Krapp's Last Tape by Beckett was first performed at the Royal Court Theatre in London.
Fernando Arrabal's Picnic on the Battlefield (Pique-nique en campagne) came out in 1958.
Genet's The Blacks (Les Nègres) was published that year but was first performed at the Théatre de Lutèce in Paris on 28 October 1959.
1959 also saw the completion of Ionesco's Rhinoceros which premiered in Paris in January 1960 at the Odeon.
Beckett's Happy Days was first performed at the Cherry Lane Theatre in New York on 17 September 1961.
Albee's Who's Afraid of Virginia Woolf? also premiered in New York the following year, on October 13.
Pinter's The Homecoming premiered in London in June 1965 at the Aldwych Theatre.
Peter Weiss's Marat/Sade (The Persecution and Assassination of Jean-Paul Marat as Performed by the Inmates of the Asylum of Charenton Under the Direction of the Marquis de Sade) was first performed in West Berlin in 1964 and in New York City a year later.
Tom Stoppard's Rosencrantz & Guildenstern Are Dead premiered at the Edinburgh Festival Fringe in 1966.
Arrabal's Automobile Graveyard (Le Cimetière des voitures) was first performed in 1966.
Lebanese author Issam Mahfouz's play The Dictator premiered in Beirut in 1969.
Beckett's Catastrophe—dedicated to then-imprisoned Czech dissident playwright Václav Havel, who became president of Czechoslovakia after the 1989 Velvet Revolution—was first performed at the Avignon Festival on July 21, 1982. The film version (Beckett on Film, 2001) was directed by David Mamet and performed by Harold Pinter, Sir John Gielgud, and Rebecca Pidgeon.
Theatrical features
Plays within this group are absurd in that they focus not on logical acts, realistic occurrences, or traditional character development; they, instead, focus on human beings trapped in an incomprehensible world subject to any occurrence, no matter how illogical. The theme of incomprehensibility is coupled with the inadequacy of language to form meaningful human connections. According to Martin Esslin, Absurdism is "the inevitable devaluation of ideals, purity, and purpose" Absurdist drama asks its viewer to "draw his own conclusions, make his own errors". Though Theatre of the Absurd may be seen as nonsense, they have something to say and can be understood". Esslin makes a distinction between the dictionary definition of absurd ("out of harmony" in the musical sense) and drama's understanding of the Absurd: "Absurd is that which is devoid of purpose... Cut off from his religious, metaphysical, and transcendental roots, man is lost; all his actions become senseless, absurd, useless".
Characters
The characters in Absurdist drama are lost and floating in an incomprehensible universe and they abandon rational devices and discursive thought because these approaches are inadequate. Many characters appear as automatons stuck in routines speaking only in cliché (Ionesco called the Old Man and Old Woman in The Chairs "übermarionettes"). Characters are frequently stereotypical, archetypal, or flat character types as in Commedia dell'arte.
The more complex characters are in crisis because the world around them is incomprehensible. Many of Pinter's plays, for example, feature characters trapped in an enclosed space menaced by some force the character can't understand. Pinter's first play was The Room – in which the main character, Rose, is menaced by Riley who invades her safe space though the actual source of menace remains a mystery – and this theme of characters in a safe space menaced by an outside force is repeated in many of his later works (perhaps most famously in The Birthday Party). In Friedrich Dürrenmatt's The Visit, the main character, Alfred, is menaced by Claire Zachanassian; Claire, richest woman in the world with a decaying body and multiple husbands throughout the play, has guaranteed a payout for anyone in the town willing to kill Alfred. Characters in Absurdist drama may also face the chaos of a world that science and logic have abandoned. Ionesco's recurring character Berenger, for example, faces a killer without motivation in The Killer, and Berenger's logical arguments fail to convince the killer that killing is wrong. In Rhinocéros, Berenger remains the only human on Earth who hasn't turned into a rhinoceros and must decide whether or not to conform. Characters may find themselves trapped in a routine, or in a metafictional conceit, trapped in a story; the title characters in Tom Stoppard's Rosencrantz & Guildenstern Are Dead, for example, find themselves in a story (Hamlet) in which the outcome has already been written.
The plots of many Absurdist plays feature characters in interdependent pairs, commonly either two males or a male and a female. Some Beckett scholars call this the "pseudocouple". The two characters may be roughly equal or have a begrudging interdependence (like Vladimir and Estragon in Waiting for Godot or the two main characters in Rosencrantz & Guildenstern Are Dead); one character may be clearly dominant and may torture the passive character (like Pozzo and Lucky in Waiting for Godot or Hamm and Clov in Endgame); the relationship of the characters may shift dramatically throughout the play (as in Ionesco's The Lesson or in many of Albee's plays, The Zoo Story for example).
Language
Despite its reputation for nonsense language, much of the dialogue in Absurdist plays is naturalistic. The moments when characters resort to nonsense language or clichés—when words appear to have lost their denotative function, thus creating misunderstanding among the characters—make the Theatre of the Absurd distinctive. Language frequently gains a certain phonetic, rhythmical, almost musical quality, opening up a wide range of often comedic playfulness. Jean Tardieu, for example, in the series of short pieces Theatre de Chambre arranged the language as one arranges music. Distinctively Absurdist language ranges from meaningless clichés to vaudeville-style word play to meaningless nonsense. The Bald Soprano, for example, was inspired by a language book in which characters would exchange empty clichés that never ultimately amounted to true communication or true connection. Likewise, the characters in The Bald Soprano—like many other Absurdist characters—go through routine dialogue full of clichés without actually communicating anything substantive or making a human connection. In other cases, the dialogue is purposefully elliptical; the language of Absurdist Theater becomes secondary to the poetry of the concrete and objectified images of the stage. Many of Beckett's plays devalue language for the sake of the striking tableau. Harold Pinter—famous for his "Pinter pause"—presents more subtly elliptical dialogue; often the primary things characters should address are replaced by ellipsis or dashes. The following exchange between Aston and Davies in The Caretaker is typical of Pinter:
ASTON. More or less exactly what you...
DAVIES. That's it … that's what I'm getting at is … I mean, what sort of jobs … (Pause.)
ASTON. Well, there's things like the stairs … and the … the bells …
DAVIES. But it'd be a matter … wouldn't it … it'd be a matter of a broom … isn't it?
Much of the dialogue in Absurdist drama (especially in Beckett's and Albee's plays, for example) reflects this kind of evasiveness and inability to make a connection. When language that is apparently nonsensical appears, it also demonstrates this disconnection. It can be used for comic effect, as in Lucky's long speech in Godot when Pozzo says Lucky is demonstrating a talent for "thinking" as other characters comically attempt to stop him:
LUCKY. Given the existence as uttered forth in the public works of Puncher and Wattmann of a personal God quaquaquaqua with white beard quaquaquaqua outside time without extension who from the heights of divine apathia divine athambia divine aphasia loves us dearly with some exceptions for reasons unknown but time will tell and suffers like the divine Miranda with those who for reasons unknown but time will tell are plunged in torment...
Nonsense may also be used abusively, as in Pinter's The Birthday Party when Goldberg and McCann torture Stanley with apparently nonsensical questions and non-sequiturs:
GOLDBERG. What do you use for pajamas?
STANLEY. Nothing.
GOLDBERG. You verminate the sheet of your birth.
MCCANN. What about the Albigensenist heresy?
GOLDBERG. Who watered the wicket in Melbourne?
MCCANN. What about the blessed Oliver Plunkett?
GOLDBERG. Speak up Webber. Why did the chicken cross the road?
As in the above examples, nonsense in Absurdist theatre may be also used to demonstrate the limits of language while questioning or parodying the determinism of science and the knowability of truth. In Ionesco's The Lesson, a professor tries to force a pupil to understand his nonsensical philology lesson:
PROFESSOR. … In Spanish: the roses of my grandmother are as yellow as my grandfather who is Asiatic; in Latin: the roses of my grandmother are as yellow as my grandfather who is Asiatic. Do you detect the difference? Translate this into … Romanian
PUPIL. The … how do you say "roses" in Romanian?
PROFESSOR. But "roses", what else? … "roses" is a translation in Oriental of the French word "roses", in Spanish "roses", do you get it? In Sardanapali, "roses"...
Plot
Traditional plot structures are rarely a consideration in the Theatre of the Absurd. Plots can consist of the absurd repetition of cliché and routine, as in Godot or The Bald Soprano. Often there is a menacing outside force that remains a mystery; in The Birthday Party, for example, Goldberg and McCann confront Stanley, torture him with absurd questions, and drag him off at the end, but it is never revealed why. In later Pinter plays, such as The Caretaker and The Homecoming, the menace is no longer entering from the outside but exists within the confined space. Other Absurdists use this kind of plot, as in Edward Albee's A Delicate Balance: Harry and Edna take refuge at the home of their friends Agnes and Tobias because they suddenly become frightened. They have difficulty explaining what has frightened them:
HARRY: There was nothing … but we were very scared.
EDNA: We … were … terrified.
HARRY: We were scared. It was like being lost: very young again, with the dark, and lost. There was no … thing … to be … frightened of, but …
EDNA: WE WERE FRIGHTENED … AND THERE WAS NOTHING.
Absence, emptiness, nothingness, and unresolved mysteries are central features in many Absurdist plots: for example, in The Chairs, an old couple welcomes a large number of guests to their home, but these guests are invisible, so all we see are empty chairs, a representation of their absence. Likewise, the action of Godot is centered around the absence of a man named Godot, for whom the characters perpetually wait. In many of Beckett's later plays, most features are stripped away and what's left is a minimalistic tableau: a woman walking slowly back and forth in Footfalls, for example, or in Breath only a junk heap on stage and the sounds of breathing.
The plot may also revolve around an unexplained metamorphosis, a supernatural change, or a shift in the laws of physics. For example, in Ionesco's Amédée, or How to Get Rid of It, a couple must deal with a corpse that is steadily growing larger and larger; Ionesco never fully reveals the identity of the corpse, how this person died, or why it's continually growing, but the corpse ultimately – and, again, without explanation – floats away. In Jean Tardieu's "The Keyhole" a lover watches a woman through a keyhole as she removes her clothes and then her flesh.
Like Pirandello, many Absurdists use meta-theatrical techniques to explore role fulfillment, fate, and the theatricality of theatre. This is true for many of Genet's plays: for example, in The Maids, two maids pretend to be their mistress; in The Balcony brothel patrons take on elevated positions in role-playing games, but the line between theatre and reality starts to blur. Another complex example of this is Rosencrantz and Guildenstern are Dead: it's a play about two minor characters in Hamlet; these characters, in turn, have various encounters with the players who perform The Mousetrap, the play-within-the-play in Hamlet. In Stoppard's Travesties, James Joyce and Tristan Tzara slip in and out of the plot of The Importance of Being Earnest.
Plots are frequently cyclical: for example, Endgame begins where the play ended – at the beginning of the play, Clov says, "Finished, it's finished, nearly finished, it must be nearly finished" – and themes of cycle, routine, and repetition are explored throughout.
Notes
Further reading
Ackerley, C. J. and S. E. Gontarski, ed. The Grove Companion to Samuel Beckett. New York: Grove P, 2004.
Adamov, Jacqueline, "Censure et représentation dans le théâtre d’Arthur Adamov", in P. Vernois (Textes recueillis et présentés par), L’Onirisme et l’insolite dans le théâtre français contemporain. Actes du colloque de Strasbourg, Paris, Editions Klincksieck, 1974.
Baker, William, and John C. Ross, comp. Harold Pinter: A Bibliographical History. London: The British Library and New Castle, DE: Oak Knoll P, 2005. (10). (13).
Bennett, Michael Y. Reassessing the Theatre of the Absurd: Camus, Beckett, Ionesco, Genet, and Pinter. New York: Palgrave Macmillan, 2011.
Bennett, Michael Y. The Cambridge Introduction to Theatre and Literature of the Absurd. Cambridge: Cambridge University Press, 2015.
Brook, Peter. The Empty Space: A Book About the Theatre: Deadly, Holy, Rough, Immediate. Touchstone, 1995. (10).
Caselli, Daniela. Beckett's Dantes: Intertextuality in the Fiction and Criticism. .
Cronin, Anthony. Samuel Beckett: The Last Modernist. New York: Da Capo P, 1997.
Driver, Tom Faw. Jean Genet. New York: Columbia UP, 1966.
Esslin, Martin. The theatre of the absurd. London: Pelican, 1980.
Gaensbauer, Deborah B. Eugène Ionesco Revisited. New York: Twayne, 1996.
Haney, W.S., II. "Beckett Out of His Mind: The Theatre of the Absurd". Studies in the Literary IMagination. Vol. 34 (2).
La Nouvelle Critique, numéro spécial "Arthur Adamov", août-septembre 1973.
Lewis, Allan. Ionesco. New York: Twayne, 1972.
McMahon, Joseph H. The Imagination of Jean Genet. New Haven: Yale UP, 1963.
Mercier, Vivian. Beckett/Beckett. Oxford UP, 1977. .
Youngberg, Q. Mommy's American Dream in Edward Albee's the American Dream. The Explicator, (2), 108.
Zhu, Jiang. "Analysis on the Artistic Features and Themes of the Theater of the Absurd". Theory & Practice in Language Studies, 3(8).
Absurdist fiction
Concepts in aesthetics
Concepts in epistemology
Concepts in metaphysics
Existentialist concepts
Metaphors
Modernist theatre
Philosophy of mind
Postmodern literature
Surrealism
Theatrical genres
Types of existentialism |
null | null | History of music | eng_Latn | The history of music covers the historical development and presence of music from prehistoric times to the present day. Though definitions of music vary wildly throughout the world, every known culture partakes in it, and music is thus considered a cultural universal. The origins of music remain highly contentious; commentators often relate it to the origin of language, with much disagreement surrounding whether music arose before, after or simultaneously with language. Many theories have been proposed by scholars from a wide range of disciplines, though none have achieved wide approval. Most cultures have their own mythical origins concerning the invention of music, generally rooted in their respective mythological, religious or philosophical beliefs.
The music of prehistoric cultures is first firmly dated to BP of the Upper Paleolithic by evidence of bone flutes, though it remains unclear whether or not the actual origins lie in the earlier Middle Paleolithic period (300,000 to 50,000 BP). There is little known about prehistoric music, with traces mainly limited to some simple flutes and percussion instruments. However, such evidence indicates that music existed to some extent in prehistoric societies such as the Xia dynasty and the Indus Valley Civilisation. Upon the development of writing, the music of literate civilizations—ancient music—was present in the major Chinese, Egyptian, Greek, Indian, Persian, Mesopotamian and Middle Eastern societies. It is difficult to make many generalizations about ancient music as a whole, but from what is known it was often characterized by monophony and improvisation. In ancient song forms, the texts were closely aligned with music, and though the oldest extant musical notation survives from this period, many texts survive without their accompanying music, such as the Rigveda and the Shijing Classic of Poetry. The eventual emergence of the Silk Road and increasing contact between cultures led to the transmission and exchange of musical ideas, practices, and instruments. Such interaction led to the Tang dynasty's music being heavily influenced by Central Asian traditions, while the Japanese gagaku court music absorbed substantial Chinese and Korean traditions.
Historically, religions have often been catalysts for music. The Vedas of Hinduism immensely influenced Indian classical music, while the Five Classics of Confucianism laid the basis for subsequent Chinese music. Following the rapid spread of Islam in the 6th century, Islamic music dominated Persia and the Arab world, and the Islamic Golden Age saw the presence of numerous important music theorists. Music written for and by the early Christian Church properly inaugurates the Western classical music tradition, which continues into medieval music where polyphony, staff notation and nascent forms of many modern instruments developed. In addition to religion or the lack thereof, a society's music is influenced by all other aspects of its culture, including social and economic organization and experience, climate, and access to technology. Many cultures have coupled music with other art forms, such as the Chinese four arts and the medieval quadrivium. The emotions and ideas that music expresses, the situations in which music is played and listened to, and the attitudes toward musicians and composers all vary between regions and periods. Many cultures have or continue to distinguish between art music (or 'classical music'), folk music and more recently, popular music.
Origins
Music is regarded as a cultural universal, though definitions of it vary wildly around the world and throughout history. As with many aspects of human cognition, it remains debated as to what extent the origins of music will ever be understood, with scholars often taking polarizing positions. The origin of music is often discussed alongside the origin of language, with the nature of their connection being the subject of much debate. However, before the mid-late 20th century, both topics were seldom given substantial attention by academics. Since the topic's resurgence, the principal source of contention is divided into three perspectives: whether music began as a kind of proto-language (a result of adaptation) that led to language; if music is a spandrel (a phenotypic byproduct of evolution) that was the result of language; or if music and language both derived from a common antecedent.
There is little consensus on any particular theory for the origin of music, which have included contributions from archaeologists, cognitive scientists, ethnomusicologists, evolutionary biologists, linguists, neuroscientists, paleoanthropologists, philosophers and psychologists (developmental and social). Some of the most prominent theories are as follows:
Music arose as an elaborate form of sexual selection, perhaps arising in mating calls. This theory, perhaps the first significant one on music's origins, is generally credited to Charles Darwin. It first appeared in Darwin's 1871 book The Descent of Man, and Selection in Relation to Sex, and has since been criticized as there is no evidence that either human sex is "more musical" thus no evidence of sexual dimorphism; there are currently no other examples of sexual selection that do not include considerable sexual dimorphism. Recent commentators, citing music's use in other animals's mating systems, have nonetheless propagated and developed Darwin's theory; such scholars include Peter J.B. Slater, Katy Payne, Björn Merker, Geoffrey Miller and Peter Todd.
Music arose alongside language, both of which supposedly descend from a "shared precursor",. The biologist Herbert Spencer was an important early proponent of this theory, as was the composer Richard Wagner, who termed the music and language's shared ancestor as "speech-music". Since the 21st-century, a number of scholars have supported this theory, particularly the archeologist Steven Mithen.
Music arose to fulfill a practical need. Propositions include:
To assist in organizing cohesive labor, first proposed by the economist Karl Bücher.
To improve the ease and range of long distance communication, first proposed by the musicologist Carl Stumpf.
To enhance communication with the divine or otherwise supernatural, first proposed by the anthropologist Siegfried Nadel.
To assist in "coordination, cohesion and cooperation", particularly in the context of families or communities.
To be a means for frightening off predators or enemies of some kind.
Music had two origins, "from speech (logogenic) and from emotional expression (pathogenic)", first proposed by the musicologist Curt Sachs. Reflecting on the diversity of music around the world, Sachs noted that some music confines to either a communicative or expressionistic form, suggesting that these aspects developed separately.
Many cultures have their own mythical origins on the creation of music. Specific figures are sometimes credited with inventing music, such as Jubal in Christian mythology, the legendary Shah Jamshid in Persian/Iranian mythology, the goddess Saraswati in Hinduism, and the muses in Ancient Greek mythology. Some cultures credit multiple originators of music; ancient Egyptian mythology associates it with numerous deities, including Amun, Hathor, Isis and Osiris, but especially Ihy. There are many stories relating to music's origins in Chinese mythology, but the most prominent is that of the musician Ling Lun, who—on the orders of the Yellow Emperor (Huangdi)—invented bamboo flute by imitating the song of the mythical fenghuang birds.
Prehistory
In the broadest sense, prehistoric music—more commonly termed primitive music in the past—encompasses all music produced in preliterate cultures (prehistory), beginning at least 6 million years ago when humans and chimpanzees last had a common ancestor. Music first arose in the Paleolithic period, though it remains unclear as to whether this was the Middle (300,000 to 50,000 BP) or Upper Paleolithic (50,000 to 12,000 BP). The vast majority of Paleolithic instruments have been found in Europe and date to the Upper Paleolithic. It is certainly possible that singing emerged far before this time, though this is essentially impossible to confirm. The potentially oldest instrument is the Divje Babe Flute from the Divje Babe cave in Slovenia, dated to 43,000 and 82,000 and made from a young cave bear femur. Purportedly used by Neanderthals, the Divje Babe Flute has received extensive scholarly attention, and whether it is truly a musical instrument or an object formed by animals is the subject of intense debate. If the former, it would be the oldest known musical instrument and evidence of a musical culture in the Middle Paleolithic. Other than the Divje Babe Flute and three other doubtful flutes, there is virtually no surviving Middle Paleolithic musical evidence of any certainty, similar to the situation in regards to visual art. The earliest objects whose designations as musical instruments are widely accepted are bone flutes from the Swabian Jura, Germany, namely from the Geissenklösterle, Hohle Fels and Vogelherd caves. Dated to the Aurignacian (of the Upper Paleolithic) and used by Early European modern humans, from all three caves there are eight examples, four made from the wing bones of birds and four from mammoth ivory; three of these are near complete. Three flutes from the Geissenklösterle are dated as the oldest, BP.
Considering the relative complexity of flutes, it is likely earlier instruments existed, akin to objects that are common in later hunter and gatherer societies, such as rattles, shakers and drums. The absence of other instruments from and before this time may be due to their use of weaker—and thus more biodegradable—materials, such as reeds, gourds, skins, and bark. A painting in the Cave of the Trois-Frères dating to BCE is thought to depict a shaman playing a musical bow.
Prehistoric cultures are thought to have had a wide variety of uses for music, with little unification between different societies. Music was likely in particular value when food and other basic needs were scarce. It is also probable that prehistoric cultures viewed music as intrinsically connected with nature, and may have believed its use influenced the natural world directly.
The earliest instruments found in prehistoric China are 12 gudi bone flutes in the modern-day Jiahu, Wuyang, Henan Province from BCE. The only instruments dated to the prehistoric Xia dynasty () are two qing, two small bells (one earthenware, one bronze) and a xun. Due to this extreme scarcity of surviving instruments and the general uncertainty surrounding most of the Xia, creating a musical narrative of the period is impractical. In the Indian subcontinent, the prehistoric Indus Valley Civilisation (from BCE in its mature state) has archeological evidence that indicates simple rattles and vessel flutes were used, while iconographical evidence suggests early harps and drums also existed. An ideogram in the later IVC contains the earliest known depiction of an arched harp, dated sometime before 1800 BCE.
Antiquity
Following the advent of writing, literate civilizations are termed part of the ancient world, a periodization which extends from the first Sumerian literature of Abu Salabikh (now Southern Iraq) of BCE, until the Post-classical era of the 6th century CE. Though the music of Ancient societies was extremely diverse, some fundamental concepts arise prominently in virtually all of them, namely monophony, improvisation and the dominance of text in musical settings. Varying song forms were present in Ancient cultures, including China, Egypt, Greece, India, Mesopotamia, Rome and the Middle East. The text, rhythm and melodies of these songs were closely aligned, as was music in general, with magic, science and religion. Complex song forms developed in later ancient societies, particularly the national festivals of China, Greece and India. Later Ancient societies also saw increased trade and transmission of musical ideas and instruments, often shepherded by the Silk Road. For example, a tuning key for a qin-zither from 4th–5th centuries BCE China includes considerable Persian iconography. In general, not enough information exists to make many other generalizations about ancient music.
The few actual examples of ancient music notation that survive usually exist on papyrus or clay tablets. Information on musical practices, genres, and thought is mainly available through literature, visual depictions, and increasingly as the period progresses, instruments. The oldest surviving written music is the Hurrian songs from Ugarit, Syria. Of these, the oldest is the Hymn to Nikkal (hymn no. 6; h. 6), which is somewhat complete and dated to BCE. However, the Seikilos epitaph is the earliest entirely complete noted musical composition. Dated to the 2nd-Century CE or later, it is an epitaph, perhaps for the wife of the unknown Seikilos.
China
Shang and Zhou
By the mid-13th century BCE, the late Shang dynasty (1600–1046 BCE) had developed writing, which mostly exists as divinatory inscriptions on the ritualistic oracle bones but also as bronze inscriptions. As many as 11 oracle script characters may refer to music to some extent, some of which could be iconographical representations of instruments themselves. The stone bells qing appears to have been particularly popular with the Shang ruling class, and while no surviving flutes have been dated to the Shang, oracle script evidence suggests they used ocarinas (xun), transverse flute (xiao and dizi), douple pipes, the mouthorgan (sheng), and maybe the pan flute (paixiao). Due to the advent of the bronze in 2000 BCE, the Shang used the material for bells—the (鈴), (鐃) and zhong (鐘)—that can be differentiated in two ways: those with or without a clapper and those struck on the inside or outside. Drums, which are not found from before the Shang, sometimes used bronze, though they were more often wooden (bangu). The aforementioned wind instruments certainly existed by the Zhou dynasty (1046–256 BCE), as did the first Chinese string instruments: the qin (or guqin) and se zithers. The Zhou saw the emergence of major court ensembles and the well known Tomb of Marquis Yi of Zeng (after 433 BCE) contains a variety of complex and decorated instruments. Of the tomb, the by-far most notable instrument is the monumental set of 65 tuned bianzhong bells, which range five octaves requiring at least five players; they are still playable and include rare inscriptions on music.
Ancient Chinese instruments served both practical and ceremonial means. People used them to appeal to supernatural forces for survival needs, while pan flutes may have been used to attract birds while hunting, and drums were common in sacrifices and military ceremonies. Chinese music has always been closely associated with dance, literature and fine arts; many early Chinese thinkers also equated music with proper morality and governance of society. Throughout the Shang and Zhou music was a symbol of power for the Imperial court, being utilized in religious services as well as the celebration of ancestors and heroes. Confucius () formally designated the music concerned with ritual and ideal morality as the superior (; "proper music"), in opposition to (; "vernacular/popular music"), which included virtually all non-ceremonial music, but particularly any that was considered excessive or lascivious. During the Warring States Period when of Confucius's lifetime, officials often ignored this distinction, preferring more lively music and using the older traditional solely for political means. Confucius and disciples such as Mencius considered this preference virtueless and saw ill of the leaders' ignorance of , a theory that held music was intrinsically connected to the universe. Thus, many aspects of Ancient Chinese music were aligned with cosmology: the 12 pitch shí-èr-lǜ system corresponded equally with certain weights and measurements; the pentatonic scale with the five ; and the eight tone classification of Chinese instruments of with the eight symbols of bagua. No actual music or texts on the performance practices of Ancient Chinese musicians survive. The Five Classics of the Zhou dynasty include musical commentary; the I Ching and Chunqiu Spring and Autumn Annals make references, while the Liji Book of Rites contains a substantial discussion (see the chapter Yue Ji Record of Music). While the Yue Jing Classic of Music is lost, the Shijing Classic of Poetry contains 160 texts to now lost songs from the Western Zhou period (1045–771).
Qin and Han
The Qin dynasty (221–206 BCE), established by Qin Shi Huang, lasted for only 15 years, but the purported burning of books resulted in a substantial loss of previous musical literature. The Qin saw the guzheng become a particularly popular instrument; as a more portable and louder zither, it meet the needs of an emerging popular music scene. During the Han dynasty (202 BCE – 220 CE), there were attempts to reconstruct the music of the Shang and Zhou, as it was now "idealized as perfect". A Music Bureau, the , was founded or at its height by at least 120 BCE under Emperor Wu of Han, and was responsible for collecting folksongs. The purpose of this was twofold; it allowed the Imperial Court to properly understand the thoughts of the common people, and it was also an opportunity for the Imperial Court to adapt and manipulate the songs to suit propaganda and political purposes. Employing ceremonial, entertainment-oriented and military musicians, the Bureau also performed at a variety of venues, wrote new music, and set music to commissioned poetry by noted figures such as Sima Xiangru. The Han dynasty had officially adopted Confucianism as the state philosophy, and the theories became a dominant philosophy. In practice, however, many officials ignored or downplayed Confucius's high regard for over music, preferring to engage in the more lively and informal later. By 7 BCE the Bureau employed 829 musicians; that year Emperor Ai either disbanded or downsized the department, due to financial limitations, and the Bureau's increasingly prominent music which conflicted with Confucianism. The Han dynasty saw a preponderance of foreign musical influences from the Middle East and Central Asia: the emerging Silk Road led to the exchange of musical instruments, and allowed travelers such as Zhang Qian to relay with new musical genres and techniques. Instruments from said cultural transmission include metal trumpets and instruments similar to the modern oboe and oud lute, the latter which became the pipa. Other preexisting instruments greatly increased in popularity, such as the qing, panpipes, and particularly the qin-zither (or guqin), which was from then on the most revered instrument, associated with good character and morality.
Greece
Greek written history extends far back into Ancient Greece, and was a major part of ancient Greek theatre. In ancient Greece, mixed-gender choruses performed for entertainment, celebration and spiritual reasons. Instruments included the most important wind instrument, the double-reed aulos, as well as the plucked string instrument, the lyre, especially the special kind called a kithara.
India
The principal sources on the music of ancient India are textual and iconographical; specifically, some theoretical treatises in Sanskrit survive, there are brief mentions in general literature and many sculptures of Ancient Indian musicians and their instruments exist. Ancient Sanskrit, Pali, and Prakrit literature frequently contains musical references, from the Vedas to the works of Kalidasa and the Ilango Adigal's epic Silappatikaram. Despite this, little is known on the actual musical practices of ancient India and the information available forces a somewhat homogeneous perspective on the music of the time, even though evidence indicates that in reality, it was far more diverse.
The monumental arts treatise Natya Shastra is among the earliest and chief sources for Ancient Indian music; the music portions alone are likely from the Gupta period (4th century to 6th century CE).
Persia and Mesopotamia
Up to the Achaemenid period
In general, it is impossible to create a thorough outline of the earliest music in Persia due to a paucity of surviving records. Evidenced by BCE Elam depictions, arched harps are the first affirmation of Persian music, though it is probable that they existed well before their artistic depictions. Elamite bull lyres from have been found in Susa, while more than 40 small Oxus trumpets have been found in Bactria and Margiana, dated to the Bactria–Margiana Archaeological Complex. The oxus trumpets seem to have had a close association with both religion and animals; a Zoroastrian myth in which Jamshid attract animals with the trumpet suggests that the Elamites used them for hunting. In many ways the earliest known musical cultures of Iran are strongly connected with those of Mesopotamia. Ancient arched harps () also exist in the latter and the scarcity of instruments makes it unclear as to which culture the harp originated. Far more bull lyres survive in Ur of Mesopotamia, notably the Bull Headed Lyre of Ur, though they are nearly identical to their contemporary Elamite counterparts. From the evidence in terracotta plaques, by the 2nd-century BCE the arched harp was displaced by the angular harps, which existed in 20-string vertical and nine-string horizontal variants. Lutes were purportedly used in Mesopotamia by at least 2300 BCE, but not until BCE do they appear in Iran, where they became the dominant string instruments of Western Iran, though the available evidence suggests its popularity was outside of the elite. The rock reliefs of Kul-e Farah show that sophisticated Persian court ensembles emerged in the 1st-century BCE, in the which the central instrument was the arched harp. The prominence of musicians in these certain rock reliefs suggests they were essential in religious ceremonies.
Like earlier periods, extremely little contemporary information on the music of the Achaemenid Empire (550–330 BCE) exists. Most knowledge on the Achaemenid musical culture comes from Greek historians. In his Histories, Herodotus noted that Achamenid priests did not use aulos music in their ceremonies, while Xenophon reflected on his visit to Persia in the Cyropaedia, mentioning the presence of many female singers at court. Athenaeus also mentions female singers when noting that 329 of them had been taken from the King of Kings Darius III by Macedonian general Parmenion. Later Persian texts assert that gōsān poet-musician minstrels were prominent and of considerable status in court.
Parthian and Sasanian periods
The Parthian Empire (247 BCE to 224 CE) saw an increase in textual and iconographical depictions of musical activity and instruments. 2nd century BCE Parthian rhuta (drinking horns) found in the ancient capital of Nisa include some of the most vivid depictions of musicians from the time. Pictorial evidence such as terracotta plaques show female harpists, while plaques from Babylon show panpipes, as well as string (harps, lutes and lyres) and percussion instruments (tambourines and clappers). Bronze statues from Dura-Europos depict larger panpipes and double aulos. Music was evidently used in ceremonies and celebrations; a Parthian-era stone frieze in Hatra shows a wedding where musicians are included, playing trumpets, tambourines, and a variety of flutes. Other textual and iconographical evidence indicates the continued prominence of gōsān minstrels. However, like the Achaemenid period, Greek writers continue to be a major source for information on Parthian music. Strabo recorded that the gōsān learned songs telling tales of gods and noblemen, while Plutarch similarly records the gōsān lauding Parthian heroes and mocking Roman ones. Plutarch also records, much to his bafflement, that rhoptra (large drums) were used by the Parthian army to prepare for war.
The Sasanian period (226–651 CE), however, has left ample evidence of music. This influx of Sasanian records suggests a prominent musical culture in the Empire, especially in the areas dominated by Zoroastrianism. Many Sassanian Shahanshahs were ardent supporters of music, including the founder of the empire Ardashir I and Bahram V. Khosrow II () was the most outstanding patron, his reign being regarded as a golden age of Persian music. Musicians in Khosrow's service include Āzādvar-e Changi, Bamshad, the harpist Nagisa (Nakisa), Ramtin, Sarkash and Barbad, who was the most famous. These musicians were usually active as minstrels, which were performers who worked as both court poets and musicians; in the Sassanian Empire there was little distinction between poetry and music.
Other Arab and African cultures
The Western African Nok culture (modern-day Nigeria) existed from BCE and left a considerable amount of sculptures. Among these are depictions of music, such as a man who shakes two objects thought to be maracas. Another sculpture includes a man with his mouth opening (possibly singing) while there is also a sculpture of a man playing a drum.
Post-classical era
Japanese gagaku music
The imperial court of Japan developed gagaku ((); ) music, originating from the Gagakuryō imperial music academy established in 701 CE during the Asuka period. Though the word gagaku derives from the Chinese yayue music, the latter originally referred to Confucian ritual music, while gagaku extends to many genres, styles and instruments. In the tradition's early history, the three main genres were wagaku (native Japanese music), sankangaku (music from the Three Kingdoms of Korea) and tōgaku (music from China's Tang dynasty), as well as more minor genres such as toragaku, gigaku, and rin’yūgaku. Uniquely among Asian music of this time, there are numerous extant scores of gagaku music from the 8th to 11th centuries.
A major shift in gagaku music occurred in the 9th century, namely the development of a distinction between tōgaku and komagaku music. Tōgaku was a Chinese-influenced style, which combined with the rin’yūgaku tradition, referred to as "Music of the Left" (sahō). Komagaku was then referred to as "Music of the Right" (uhō), encompassing music influenced by both Korea (sankangaku) and Balhae (bokkaigaku). Though this division was prominent, it was not strict and the tōgaku and komagaku styles nonetheless interlaced and influenced each other. The long Heian period (794–1185) saw much patronage of gagaku music from the court, as it accompanied many festivals and celebrations. Numerous new genres emerged at this time, such as the saibara and rōei song forms. Gagaku ensembles consist of a wide variety of instruments and are the largest such formations in traditional Japanese music.
Medieval Europe
Modern scholars generally define 'Medieval music' as the music of Western Europe during the Middle Ages, from approximately the 6th to 15th centuries. Music was certainly prominent in the Early Middle Ages, as attested by artistic depictions of instruments, writings about music, and other records; however, the only repertory of music which has survived from before 800 to the present day is the plainsong liturgical music of the Roman Catholic Church, the largest part of which is called Gregorian chant. Pope Gregory I, who gave his name to the musical repertory and may himself have been a composer, is usually claimed to be the originator of the musical portion of the liturgy in its present form, though the sources giving details on his contribution date from more than a hundred years after his death. Many scholars believe that his reputation has been exaggerated by legend. Most of the chant repertory was composed anonymously in the centuries between the time of Gregory and Charlemagne.
During the 9th century, several important developments took place. First, there was a major effort by the Church to unify the many chant traditions and suppress many of them in favor of the Gregorian liturgy. Second, the earliest polyphonic music was sung, a form of parallel singing known as organum. Third, and of the greatest significance for music history, notation was reinvented after a lapse of about five hundred years, though it would be several more centuries before a system of pitch and rhythm notation evolved having the precision and flexibility that modern musicians take for granted.
Several schools of polyphony flourished in the period after 1100: the St. Martial school of organum, the music of which was often characterized by a swiftly moving part over a single sustained line; the Notre Dame school of polyphony, which included the composers Léonin and Pérotin, and which produced the first music for more than two parts around 1200; the musical melting-pot of Santiago de Compostela in Galicia, a pilgrimage destination and site where musicians from many traditions came together in the late Middle Ages, the music of whom survives in the Codex Calixtinus; and the English school, the music of which survives in the Worcester Fragments and the Old Hall Manuscript. Alongside these schools of sacred music a vibrant tradition of the secular song developed, as exemplified in the music of the troubadours, trouvères, and Minnesänger. Much of the later secular music of the early Renaissance evolved from the forms, ideas, and the musical aesthetic of the troubadours, courtly poets, and itinerant musicians, whose culture was largely exterminated during the Albigensian Crusade in the early 13th century.
Forms of sacred music which developed during the late 13th century included the motet, conductus, discant, and clausulae. One unusual development was the Geisslerlieder, the music of wandering bands of flagellants during two periods: the middle of the 13th century (until they were suppressed by the Church); and the period during and immediately following the Black Death, around 1350, when their activities were vividly recorded and well-documented with notated music. Their music mixed folk song styles with penitential or apocalyptic texts. The 14th century in European music history is dominated by the style of the ars nova, which by convention is grouped with the medieval era in music, even though it had much in common with early Renaissance ideals and aesthetics. Much of the surviving music of the time is secular, and tends to use the formes fixes: the ballade, the virelai, the lai, the rondeau, which correspond to poetic forms of the same names. Most pieces in these forms are for one to three voices, likely with instrumental accompaniment: famous composers include Guillaume de Machaut and Francesco Landini.
Byzantine
Prominent and diverse musical practices were present in the Byzantine Empire, which existed by 395 to 1453. Both sacred and secular music were commonplace, with sacred music frequently used in church services and secular music in many events including, ceremonies dramas, ballets, banquets, festivals and sports games. However, despite its popularity, secular Byzantine music was harshly criticized by the Church Fathers. Composers of sacred music, especially hymns and chants, are generally well documented throughout the history of Byzantine music. However, those before the reign of Justinian I are virtually unknown; the monks Anthimos, Auxentios and Timokles are said to have written troparia, but only the text to a single one by Auxentios survives. The first major form was the kontakion, of which Romanos the Melodist was the foremost composer. In the late 7th century the kanōn overtook the kontakion in popularity; Andrew of Crete became its first significant composer, and is traditionally credited as the genre's originator, though modern scholars now doubt this attribution. The kañon reached its peak with the music of John of Damascus and Cosmas of Maiuma and later Theodore of Stoudios and Theophanes the Branded in the 8th and 9th centuries respectively. Composers of secular music are considerably less documented. Not until late in the empire's history are composers known by name, with Joannes Koukouzeles, Xenos Korones and Joannes Glykys as the leading figures.
Like their Western counterparts of the same period, Byzantine composers were primarily men. Kassia is a major exception to this; she was a prolific and important composer of sticheron hymns and the only woman whose works entered the Byzantine liturgy. A few other women are known to have been composers, Thekla, Theodosia, Martha and the daughter of John Kladas (her given name is unrecorded). Only the latter has any surviving work, a single antiphon. Some Byzantine emperors are known to have been composers, such as Leo VI the Wise and Constantine VII.
Early modern and modern periods
Indian classical music
During the ancient and medieval periods, the classical music of the Indian subcontinent was a largely unified practice. By the 14th century, socio-political turmoil inaugurated by the Delhi Sultanate began to isolate Northern and Southern India, and independent traditions in each region began emerging. By the 16th-century two distinct styles had formed: the Hindustani classical music of the North and the Carnatic classical music of the South. One of the major differences between them is that the Northern Hindustani vein was considerably influenced by the Persian and Arab musical practices of the time. Carnatic music is largely devotional; the majority of the songs are addressed to the Hindu deities.
Indian classical music (marga) is monophonic and based on a single melody line or raga rhythmically organized through talas.
Western classical music
Renaissance
The beginning of the Renaissance in music is not as clearly marked as the beginning of the Renaissance in the other arts, and unlike in the other arts, it did not begin in Italy, but in northern Europe, specifically in the area currently comprising central and northern France, the Netherlands, and Belgium. The style of the Burgundian composers, as the first generation of the Franco-Flemish school, is known, was at first a reaction against the excessive complexity and mannered style of the late 14th century ars subtilior, and contained clear, singable melody and balanced polyphony in all voices. The most famous composers of the Burgundian school in the mid-15th century are Guillaume Dufay, Gilles Binchois, and Antoine Busnois.
By the middle of the 15th century, composers and singers from the Low Countries and adjacent areas began to spread across Europe, especially into Italy, where they were employed by the papal chapel and the aristocratic patrons of the arts (such as the Medici, the Este, and the Sforza families). They carried their style with them: smooth polyphony which could be adapted for sacred or secular use as appropriate. Principal forms of sacred musical composition at the time were the mass, the motet, and the laude; secular forms included the chanson, the frottola, and later the madrigal.
The invention of printing had an immense influence on the dissemination of musical styles, and along with the movement of the Franco-Flemish musicians, contributed to the establishment of the first truly international style in European music since the unification of Gregorian chant under Charlemagne. Composers of the middle generation of the Franco-Flemish school included Johannes Ockeghem, who wrote music in a contrapuntally complex style, with varied texture and an elaborate use of canonical devices; Jacob Obrecht, one of the most famous composers of masses in the last decades of the 15th century; and Josquin des Prez, probably the most famous composer in Europe before Palestrina, and who during the 16th century was renowned as one of the greatest artists in any form. Music in the generation after Josquin explored increasing complexity of counterpoint; possibly the most extreme expression is in the music of Nicolas Gombert, whose contrapuntal complexities influenced early instrumental music, such as the canzona and the ricercar, ultimately culminating in Baroque fugal forms.
By the middle of the 16th century, the international style began to break down, and several highly diverse stylistic trends became evident: a trend towards simplicity in sacred music, as directed by the Counter-Reformation Council of Trent, exemplified in the music of Giovanni Pierluigi da Palestrina; a trend towards complexity and chromaticism in the madrigal, which reached its extreme expression in the avant-garde style of the Ferrara School of Luzzaschi and the late century madrigalist Carlo Gesualdo; and the grandiose, sonorous music of the Venetian school, which used the architecture of the Basilica San Marco di Venezia to create antiphonal contrasts. The music of the Venetian school included the development of orchestration, ornamented instrumental parts, and continuo bass parts, all of which occurred within a span of several decades around 1600. Famous composers in Venice included the Gabrielis, Andrea and Giovanni, as well as Claudio Monteverdi, one of the most significant innovators at the end of the era.
Most parts of Europe had active and well-differentiated musical traditions by late in the century. In England, composers such as Thomas Tallis and William Byrd wrote sacred music in a style similar to that written on the continent, while an active group of home-grown madrigalists adapted the Italian form for English tastes: famous composers included Thomas Morley, John Wilbye and Thomas Weelkes. Spain developed instrumental and vocal styles of its own, with Tomás Luis de Victoria writing refined music similar to that of Palestrina, and numerous other composers writing for the new guitar. Germany cultivated polyphonic forms built on the Protestant chorales, which replaced the Roman Catholic Gregorian Chant as a basis for sacred music, and imported the style of the Venetian school (the appearance of which defined the start of the Baroque era there). In addition, German composers wrote enormous amounts of organ music, establishing the basis for the later Baroque organ style which culminated in the work of J.S. Bach. France developed a unique style of musical diction known as musique mesurée, used in secular chansons, with composers such as Guillaume Costeley and Claude Le Jeune prominent in the movement.
One of the most revolutionary movements in the era took place in Florence in the 1570s and 1580s, with the work of the Florentine Camerata, who ironically had a reactionary intent: dissatisfied with what they saw as contemporary musical depravities, their goal was to restore the music of the ancient Greeks. Chief among them were Vincenzo Galilei, the father of the astronomer, and Giulio Caccini. The fruits of their labors was a declamatory melodic singing style known as monody, and a corresponding staged dramatic form: a form known today as opera. The first operas, written around 1600, also define the end of the Renaissance and the beginning of the Baroque eras.
Music prior to 1600 was modal rather than tonal. Several theoretical developments late in the 16th century, such as the writings on scales on modes by Gioseffo Zarlino and Franchinus Gaffurius, led directly to the development of common practice tonality. The major and minor scales began to predominate over the old church modes, a feature which was at first most obvious at cadential points in compositions, but gradually became pervasive. Music after 1600, beginning with the tonal music of the Baroque era, is often referred to as belonging to the common practice period.
Baroque
The Baroque era took place from 1600 to 1750, as the Baroque artistic style flourished across Europe and, during this time, music expanded in its range and complexity. Baroque music began when the first operas (dramatic solo vocal music accompanied by orchestra) were written. During the Baroque era, polyphonic contrapuntal music, in which multiple, simultaneous independent melody lines were used, remained important (counterpoint was important in the vocal music of the Medieval era). German, Italian, French, Dutch, Polish, Spanish, Portuguese, and English Baroque composers wrote for small ensembles including strings, brass, and woodwinds, as well as for choirs and keyboard instruments such as pipe organ, harpsichord, and clavichord. During this period several major music forms were defined that lasted into later periods when they were expanded and evolved further, including the fugue, the invention, the sonata, and the concerto. The late Baroque style was polyphonically complex and richly ornamented. Important composers from the Baroque era include Johann Sebastian Bach, Arcangelo Corelli, François Couperin, Girolamo Frescobaldi, George Frideric Handel, Jean-Baptiste Lully, Claudio Monteverdi, Georg Philipp Telemann and Antonio Vivaldi.
Classical
The music of the Classical period is characterized by homophonic texture, or an obvious melody with accompaniment. These new melodies tended to be almost voice-like and singable, allowing composers to actually replace singers as the focus of the music. Instrumental music therefore quickly replaced opera and other sung forms (such as oratorio) as the favorite of the musical audience and the epitome of great composition. However, opera did not disappear: during the classical period, several composers began producing operas for the general public in their native languages (previous operas were generally in Italian).
Along with the gradual displacement of the voice in favor of stronger, clearer melodies, counterpoint also typically became a decorative flourish, often used near the end of a work or for a single movement. In its stead, simple patterns, such as arpeggios and, in piano music, Alberti bass (an accompaniment with a repeated pattern typically in the left hand), were used to liven the movement of the piece without creating a confusing additional voice. The now-popular instrumental music was dominated by several well-defined forms: the sonata, the symphony, and the concerto, though none of these were specifically defined or taught at the time as they are now in music theory. All three derive from sonata form, which is both the overlying form of an entire work and the structure of a single movement. Sonata form matured during the Classical era to become the primary form of instrumental compositions throughout the 19th century.
The early Classical period was ushered in by the Mannheim School, which included such composers as Johann Stamitz, Franz Xaver Richter, Carl Stamitz, and Christian Cannabich. It exerted a profound influence on Joseph Haydn and, through him, on all subsequent European music. Wolfgang Amadeus Mozart was the central figure of the Classical period, and his phenomenal and varied output in all genres defines our perception of the period. Ludwig van Beethoven and Franz Schubert were transitional composers, leading into the Romantic period, with their expansion of existing genres, forms, and even functions of music.
Romantic
In the Romantic period, music became more expressive and emotional, expanding to encompass literature, art, and philosophy. Famous early Romantic composers include Schumann, Chopin, Mendelssohn, Bellini, Donizetti, and Berlioz. The late 19th century saw a dramatic expansion in the size of the orchestra, and in the role of concerts as part of urban society. Famous composers from the second half of the century include Johann Strauss II, Brahms, Liszt, Tchaikovsky, Verdi, and Wagner. Between 1890 and 1910, a third wave of composers including Grieg, Dvořák, Mahler, Richard Strauss, Puccini, and Sibelius built on the work of middle Romantic composers to create even more complex – and often much longer – musical works. A prominent mark of late 19th-century music is its nationalistic fervor, as exemplified by such figures as Dvořák, Sibelius, and Grieg. Other prominent late-century figures include Saint-Saëns, Fauré, Rachmaninoff, Franck, Debussy and Rimsky-Korsakov.
20th and 21st-century
The 20th century saw a revolution in music listening as the radio gained popularity worldwide and new media and technologies were developed to record, edit and distribute music. Music performances became increasingly visual with the broadcast and recording of performances.
20th-century music brought new freedom and wide experimentation with new musical styles and forms that challenged the accepted rules of music of earlier periods. The invention of musical amplification and electronic instruments, especially the synthesizer, in the mid-20th century revolutionized classical and popular music, and accelerated the development of new forms of music.
As for classical music, two fundamental schools determined the course of the century: that of Arnold Schoenberg and that of Igor Stravinsky. However, other composers also had a notable influence. For example: Béla Bartók, Anton Webern, Dmitri Shostakovich, Olivier Messiaen, John Cage, Benjamin Britten, Karlheinz Stockhausen, Sofia Gubaidulina, Krzysztof Penderecki, Brian Ferneyhough, Kaija Saariaho.
The 20th century saw the unprecedented dissemination of popular music, that is, music with a wide appeal. The term has its roots in the music of the American Tin Pan Alley, a group of prominent musicians and publishers who began to emerge during the 1880s in New York City. Although popular music sometimes is known as "pop music", the two terms are not interchangeable. Popular music is a generic term for a wide variety of genres of music that appeal to the tastes of a large segment of the population, whereas pop music usually refers to a specific musical genre within popular music. Popular music songs and pieces typically have easily singable melodies. The song structure of popular music commonly involves repetition of sections, with the verse and chorus or refrain repeating throughout the song and the bridge providing a contrasting and transitional section within a piece.
Notes
References
Sources
Books and chapters
(In )
(In )
(In )
(In )
(In )
(In
(In )
(In
(In
(In
(In )
(In )
(In )
(In )
(In )
(In )
Journal and encyclopedia articles
(In )
(In )
(In )
(In
(In
(In
(In )
(In )
(In )
(In )
External links
Music History from Primary Sources from the Moldenhauer Archives of the Library of Congress
All ten volumes of the Garland Encyclopedia of World Music |
null | null | Sugar Ray Leonard | eng_Latn | Ray Charles Leonard (born May 17, 1956), best known as "Sugar" Ray Leonard, is an American former professional boxer, motivational speaker, and occasional actor. Often regarded as one of the greatest boxers of all time, he competed between 1977 and 1997, winning world titles in five weight divisions; the lineal championship in three weight divisions; as well as the undisputed welterweight title. Leonard was part of "The Four Kings", a group of boxers who all fought each other throughout the 1980s, consisting of Leonard, Roberto Durán, Thomas Hearns, and Marvin Hagler.
"The Fabulous Four" created a wave of popularity in the lower weight classes that kept boxing relevant in the post-Muhammad Ali era, during which Leonard defeated future fellow International Boxing Hall of Fame inductees Hearns, Durán, Hagler, and Wilfred Benítez. Leonard was also the first boxer to earn more than $100 million in purses, and was named "Boxer of the Decade" in the 1980s. The Ring magazine named him Fighter of the Year in 1979 and 1981, while the Boxing Writers Association of America (BWAA) named him Fighter of the Year in 1976, 1979, and 1981. In 2002, Leonard was voted by The Ring as the ninth greatest fighter of the last 80 years; In 2016, he was voted by The Ring to be the greatest living fighter; and BoxRec ranks him as the 14th greatest boxer of all time, pound for pound.
Early life
Leonard, the fifth of seven children of Cicero and Getha Leonard, was born in Wilmington, North Carolina. He was named after Ray Charles, his mother's favorite singer. The family moved to Washington, D.C., when he was three, and they settled permanently in Palmer Park, Maryland when he was ten. His father worked as a supermarket night manager and his mother was a nurse. He attended Parkdale High School. Leonard was a shy child, and aside from the time he nearly drowned in a creek during a flood in Seat Pleasant, Maryland, his childhood was uneventful. He stayed home a lot, reading comic books and playing with his dog. His mother said: "He never did talk too much. We never could tell what he was thinking. But I never had any problems with him. I never had to go to school once because of him."
Amateur career
Leonard started boxing at the Palmer Park Recreation Center in 1969. His older brother, Roger, started boxing first. Roger helped start the boxing program, urging the center's director, Ollie Dunlap, to form a team. Dave Jacobs, a former boxer, and Janks Morton volunteered as boxing coaches. Roger won some trophies and showed them off in front of Ray, goading him to start boxing.
In 1972, Leonard boxed in the featherweight quarterfinals of the National AAU Tournament, losing by decision to Jerome Artis. It was his first defeat. Later that year, he boxed in the Eastern Olympic Trials. The rules stated that a boxer had to be seventeen to box in international competition, so Leonard, only sixteen, lied about his age. He made it to the lightweight semifinals, losing a disputed decision to Greg Whaley, who took such a beating that he wasn't allowed to continue in the trials and never boxed again.
Sarge Johnson, assistant coach of the US Olympic Boxing Team, said to Dave Jacobs, "That kid you got is sweet as sugar". The nickname stuck. However, given his style and first name, it was probably only a matter of time before people started calling him Sugar Ray, after the man many consider to be the best boxer of all time, Sugar Ray Robinson.
In 1973, Leonard won the National Golden Gloves Lightweight Championship, but lost to Randy Shields in the lightweight final of the National AAU Tournament. The following year, Leonard won the National Golden Gloves and National AAU Lightweight Championships. Leonard suffered his last two losses as an amateur in 1974. He lost a disputed decision to Anatoli Kamnev in Moscow, after which, Kamnev gave the winner's trophy to Leonard. In Poland, local fighter Jan Kwacz was given a disqualification victory over Leonard after being knocked down three times in the first round but the referee ruled that Leonard had punched after the bell.
Leonard won the National Golden Gloves and National AAU Light Welterweight Championships in 1974. The following year, he again won the National AAU Light Welterweight Championship, as well as the Light Welterweight Championship at the Pan American Games.
In 1976, Leonard made the U.S. Olympic Team as the light welterweight representative. The team also included Leon and Michael Spinks, Howard Davis Jr., Leo Randolph, Charles Mooney, and John Tate. Many consider the 1976 U.S. team to be the greatest boxing team in the history of the Olympics. Leonard won his first four Olympic bouts by 5–0 decisions. In the semifinals, he faced Kazimierz Szczerba and won a 5–0 decision.
In the final, Leonard boxed the great Cuban knockout artist Andrés Aldama, who scored five straight knockouts to reach the final. Leonard landed several good left hooks in the first round. In the second, he dropped Aldama with a left to the chin. Late in the final round, he again hurt Aldama, which brought a standing eight count from the referee.
With only a few seconds left in the fight, a Leonard combination forced another standing eight count. Leonard was awarded a 5–0 decision and the Olympic gold medal. Afterward, Leonard announced, "I'm finished...I've fought my last fight. My journey has ended, my dream is fulfilled. Now I want to go to school." He was given a scholarship to the University of Maryland, a gift from the citizens of Glenarden, Maryland. He planned to study business administration and communications. He finished his amateur career with a record of 165–5 and 75 KOs.
Achievements
1973 National Golden Gloves Lightweight Champion, defeating Hilmer Kenty
1973 National AAU Light Welterweight Championship runner-up, losing to Randy Shields
1974 National Golden Gloves Light Welterweight Champion, defeating Jeff Lemeir
1974 National AAU Light Welterweight Champion, defeating Paul Sherry
1974 North American Championships Gold Medalist, defeating Robert Proulx
1975 National AAU Light Welterweight Champion, defeating Milton Seward
1975 North American Championships Gold Medalist, defeating Michel Briere
1975 Pan American Games Light Welterweight Gold Medalist, defeating Victor Corona
1976 Olympic Light Welterweight Gold Medalist, defeating Andrés Aldama
Olympic results
1/32: Defeated Ulf Carlsson (Sweden) by unanimous decision, 5–0
1/16: Defeated Valery Limasov (Soviet Union) by unanimous decision, 5–0
1/8: Defeated Clinton McKenzie (Great Britain) by unanimous decision, 5–0
1/4: Defeated Ulrich Beyer (East Germany) by unanimous decision, 5–0
1/2: Defeated Kazimierz Szczerba (Poland) by unanimous decision, 5–0
Finals: Defeated Andrés Aldama (Cuba) by unanimous decision, 5–0
Change of plans
Juanita Wilkinson, Leonard's high school girlfriend, told him she was pregnant in the summer of 1973. They decided to have the baby but marriage would be put off until after the Olympics in 1976. Leonard would continue to pursue his Olympic dream while she and the baby, Ray Charles Leonard Jr., lived with her parents. When Leonard boxed in the Olympics, he had a picture of Wilkinson taped to his sock.
Shortly before the Olympics, Wilkinson had filed an application to receive $156 a month in child support payments from Prince George's County, Maryland. She named Leonard as the father and the county's state attorney's office filed a civil suit against Leonard to establish paternity and get support payments for the child. Leonard learned of the suit several days after returning home from the Olympics. The headline in the Washington Star read, "Sugar Ray Leonard Named in Welfare Dept. Paternity Suit".
Wilkinson went to the Olympics to watch Leonard box, but she did not tell him about the suit and never asked him for any money. "I didn't feel like being bothered by all those complications by asking him for any money for support", she said. Leonard pledged he would support his son, even if he had to scrap plans to attend college.
Leonard had hoped to get lucrative endorsements following his gold medal win, but the negative publicity from the paternity suit chased off any big commercial possibilities. To make matters worse, his father was hospitalized with meningitis and his mother suffered a heart attack. With neither parent able to work, with his child and the mother of his child to support, and without any endorsement opportunities, Leonard decided to become a professional boxer.
Professional career
Early professional career
When Leonard decided to turn professional, Janks Morton introduced him to Mike Trainer, a friend of his who was an attorney. Trainer talked 24 of his friends and clients into underwriting Leonard's career with an investment of $21,000 to be repaid within four years at 8% interest. Trainer then made Leonard the sole stockholder in Sugar Ray Leonard, Inc. Angelo Dundee, Muhammad Ali's trainer, was brought in to be Leonard's trainer and manager. Many of the people being considered wanted absolute control and a cut somewhere near the manager's traditional 33%. Dundee had a different proposition. Although he would prescribe the training procedures, he would leave the day-to-day work to Dave Jacobs and Janks Morton. He would also choose Leonard's opponents. For his services, Dundee would get 15% of Leonard's purse.
Leonard made his professional debut on February 5, 1977, before a crowd of 10,270 at the Civic Center in Baltimore. He was paid $40,044 for the fight. His opponent was Luis "The Bull" Vega, whom he defeated by a six-round unanimous decision. After the fight, Leonard paid back his $21,000 loan to the investors.
In his fourteenth professional fight, Leonard fought his first world-ranked opponent, Floyd Mayweather, who was ranked seventeenth. The fight took place on September 9, 1978. Leonard won by a tenth-round knockout. A month later, Leonard defeated his old amateur nemesis Randy Shields by a ten-round unanimous decision.
On August 12, 1979, Leonard knocked out Pete Ranzany in four rounds to win the NABF Welterweight Championship.
The following month, he made his first title defense against Andy Price. Price, an up-and-coming contender who was sponsored by Marvin Gaye, had a reputation for prolonged bouts in earlier fights and was believed by sports reporters to defeat or give a long fight to Leonard. Although Price landed multiple good blows, Leonard knocked him out in the first round, advancing his record to 25–0 with 16 knockouts.
First world title
Leonard vs. Benitez
Leonard fought Wilfred Benítez for the WBC Welterweight Championship on November 30, 1979, at Caesar's Palace in Las Vegas, Nevada. There was a capacity crowd of about 4,600. Leonard received $1 million and Benitez, a two-division champion with a record of 38–0–1, received $1.2 million.
It was a highly competitive and tactical battle. In the first round, Leonard rocked Benitez with a left hook that came off a jab and right cross. Late in the third, Leonard dropped Benitez on the seat of his pants with a stiff left jab. More embarrassed than hurt, Benitez got up quickly. Benitez started improving in the fourth, slipping numerous punches and finding the range with his right hand. "I wasn't aware I was in a championship early because I hit him so easy", Leonard said. "But then he adjusted to my style. It was like looking in a mirror".
In the sixth, there was an accidental clash of heads, which opened a cut on the forehead of Benitez. Blood flowed down his forehead and the bridge of his nose but stayed out of his eyes.
Leonard landed the harder punches and had Benitez hurt several times late in the fight, but Leonard couldn't put him away. Benitez was very slick. "No one, I mean no one, can make me miss punches like that", Leonard said.
Going into the final round, Leonard led by scores of 137–130, 137–133, and 136–134. The two went toe-to-toe in the fifteenth. Late in the round, Leonard dropped Benitez with a left. He got up, but after a few more punches, the referee stopped the fight. The time was 2:54 of round fifteen.
The Boxing Writers Association of America and The Ring named Leonard "Fighter of the Year" for 1979.
Leonard vs. Green
Leonard made his first title defense in Landover, Maryland, on March 31, 1980. His opponent was Dave "Boy" Green. The British challenger had a record of 33–2. In the fourth round, Leonard knocked Green out with a devastating left hook. Leonard called it "the hardest single punch I ever threw."
The Brawl in Montreal
On June 20, 1980, Leonard returned to the Olympic Stadium in Montreal to defend his title against Roberto Durán before a crowd of 46,317. Durán, the former Undisputed World Lightweight Champion for years, had a record of 71–1 and was the #1 welterweight contender and considered the best "Pound for Pound" fighter in the world. Durán received $1.5 million and Leonard, working for a percentage of the closed-circuit gate as well as a guarantee, received over $9 million.
Angelo Dundee counseled Leonard to box, to move side to side and not to get caught on the ropes. However, Leonard decided to fight Durán's way. "Flat-footed", he said. "I will not run."
Durán forced the issue and took the fight to Leonard, cutting off the ring and denying Leonard space to fight his fight. Durán attacked at almost every turn. Leonard battled back again and again, but he had to work just to find room to breathe and swing, at times simply to survive. In the second, Durán rocked Leonard with a left hook, sending him into the ropes. Leonard started to do better by the fifth round, finding some punching room and throwing numerous multi-punch combinations. The two fought with great intensity throughout the fight. According to Bill Nack:
It was, from almost the opening salvo, a fight that belonged to Durán. The Panamanian seized the evening and gave it what shape and momentum it had. He took control, attacking and driving Leonard against the ropes, bulling him back, hitting him with lefts and rights to the body as he maneuvered the champion against the ropes from corner to corner. Always moving forward, he mauled and wrestled Leonard, scoring inside with hooks and rights. For three rounds Durán drove at Sugar Ray with a fury, and there were moments when it seemed the fight could not last five. Unable to get away, unable to counter and unable to slide away to open up the ring, Leonard seemed almost helpless under the assault. Now and then he got loose and countered—left-right-left to Durán's bobbing head—but he missed punches and could not work inside, could not jab, could not mount an offense to keep Durán at bay.
Durán was awarded a unanimous decision, although it was mistakenly read as a majority decision in the ring. The scorecard of judge Angelo Poletti was incorrectly added and announced as 147–147. He actually scored it 148–147. In rounds, he had it three for Durán, two for Leonard, and ten even. Sports Illustrated called his scorecard "a monument to indecision." Judges Raymond Baldeyrou and Harry Gibbs scored the fight 146–144 and 145–144, respectively. Associated Press had it 144–141 for Durán, while The New York Times had Leonard ahead 144–142.
"I did the best I could", Leonard said. "I think I pretty much fought from the heart." Asked if Leonard was the best he ever fought, Durán thought for a moment and then answered, "Si, si." Durán said. "He does have a heart. That's why he's living."
"No Más" in New Orleans
The rematch, billed as "Stone vs. Sugar.. Once Again", took place November 25, 1980 at the Louisiana Superdome in New Orleans in front of 25,038 fans. Leonard received $7 million and Durán received $8 million.
Dave Jacobs disagreed with the decision to have an immediate rematch with Durán and terminated his relationship with Leonard when the rematch was made. "My idea is that he should have a tuneup fight before he fights with Roberto again", Jacobs said. "I think he won the fight with Durán, but I don't think it is healthy for him to be fighting Durán right away".
After the Montreal fight Durán went on a partying binge and ballooned in weight. Leonard was aware of this, and in an interview for Beyond the Glory he said: "My intention was to fight Durán ASAP because I knew Durán's habits. I knew he would indulge himself, he'd gain 40–50 lbs and then sweat it off to make 147." Unlike the fight in Montreal, Leonard used his superior speed and movement to outbox and befuddle Durán. "The whole fight, I was moving, I was moving", Leonard said. "And Voom! I snapped his head back with a jab. Voom! I snapped it back again. He tried to get me against the ropes, I'd pivot, spin off and Pow! Come under with a punch."
In round seven, Leonard started to taunt Durán. Leonard's most memorable punch came late in the round. Winding up his right hand, as if to throw a bolo punch, Leonard snapped out a left jab and caught Durán flush in the face. "It made his eyes water", Leonard said. He continued to taunt Durán mercilessly. He stuck out his chin, inviting Durán to hit it. Durán hesitated. Leonard kept it up, continuing to move, stop, and mug.
In the closing seconds of the eighth round, Durán turned his back to Leonard and quit, saying to referee Octavio Meyran, "No Más" (English: "No more"). Leonard was the winner by a technical knockout at 2:44 of round eight, regaining the WBC Welterweight Championship. Leonard led by scores of 68–66, 68–66 and 67–66.
Durán said he quit because of stomach cramps, caused by overeating after the weigh-in. "At the end of the fifth round, I got cramps in my stomach and it kept getting worse and worse", Duran later said. "I felt weaker and weaker in my body and arms." He then announced, "I am retiring from boxing right now." During the night Durán was admitted to a hospital with stomach pains, and discharged the following day.
Everyone was surprised by Durán's actions, none more so than his veteran trainers, Freddie Brown and Ray Arcel. "I was shocked", Brown said. "There was no indication that he was in pain or getting weak." Arcel was angry. "That's it", he said. "I've had it. This is terrible. I've handled thousands of fighters and never had anyone quit on me. I think he needs a psychiatrist more than he needs anything else." Durán's manager, Carlos Eleta, said, "Durán didn't quit because of stomach cramps. He quit because he was embarrassed. I know this." According to Randy Gordon, who witnessed Durán's antics beforehand and was in his dressing room immediately afterwards, Durán quit because of his huge eating binge prior to the fight.
"I made him quit", Leonard said. "To make a man quit, to make Roberto Durán quit, was better than knocking him out."
Second world title
Leonard vs. Bonds
On March 28, 1981, Leonard defended his title against Larry Bonds, the WBC sixth-ranked contender, at the Carrier Dome in Syracuse, New York. Bonds was a southpaw, which made him a good opponent for Leonard, given that his next opponent was scheduled to be the WBA Light Middleweight Champion Ayub Kalule, a southpaw.
Leonard was the aggressor throughout, with Bonds circling the ring. He staggered Bonds with a right in the fourth round and dropped him with a follow-up combination. Bonds got up and continued to move, with Leonard in pursuit. Leonard dropped him again in the tenth. Bonds rose but Leonard didn't let him off the hook. The referee stopped the fight with Bonds taking punishment in a corner.
Leonard vs. Kalule
Leonard moved up to the junior middleweight division and faced Kalule on June 25, 1981, at the Astrodome in Houston, Texas. Kalule, who was 36–0, had been the WBA Light Middleweight Champion for two years.
Kalule and his handlers had expected Leonard to use lateral movement against him, but Leonard chose to fight inside instead. After eight tough rounds, Leonard was ahead although Kalule appeared to be coming on strong in the eight and ninth. Leonard finally hurt him with a right to the head. Shortly afterward, Leonard dropped him with a flurry of punches. Kalule got up but the referee waved it off. Leonard celebrated his victory with a full 360-degree, no-hands flip. Despite an official stoppage time of 2.59, the fight was actually stopped at 3.06 into the round, meaning Kalule should have been saved by the bell.
The Showdown
Promoted as "The Showdown", Leonard fought Thomas Hearns on September 16, 1981, at Caesars Palace in Las Vegas to unify the World Welterweight Championship in a scheduled fifteen-rounder. They fought before a live crowd of 23,618. Hearns was paid $5.1 million, and Leonard made over $11 million. The fight grossed over $35 million. The live gate was $5.9 million, and the revenue from pay-per-view was $7.5 million.
Hearns, 32–0 with 30 knockouts, won the WBA Welterweight Championship in 1980, scoring a second-round knockout of José "Pipino" Cuevas in Detroit, Michigan. He made three successful title defenses, stopping Luis Primera, Randy Shields, and Pablo Baez.
The fight began as expected, Leonard boxing from a distance and Hearns stalking. Leonard had difficulty with Hearns' long reach and sharp jab. By the end of round five, Leonard had a growing swelling under his left eye, and Hearns had built a considerable lead on the scorecards. Leonard, becoming more aggressive, hurt Hearns in the sixth with a left hook to the chin. Leonard battered Hearns in rounds six and seven, but Hearns regrouped. Hearns started to stick and move, and he started to pile up points again. The roles reversed: Leonard became the stalker and Hearns became the boxer. The fight billed as a classic showdown between a powerful knockout artist and the best boxer/puncher the welterweight division had seen in decades devolved into a slow, tactical fight.
Hearns won rounds nine through twelve on all three scorecards. Between rounds twelve and thirteen, Angelo Dundee told Leonard, "You're blowing it, son! You're blowing it!".
Leonard, with a badly swollen left eye, came out roaring for the thirteenth round. After hurting Hearns with a right, Leonard exploded with a combination of punches. Hearns' legs were clearly gone and after more pressure from Leonard he was bundled through the ropes, no knockdown was given as it wasn't a punch that sent him there. Hearns managed to rise, but was dropped by a flurry of hard punches near the end of the round.
In round fourteen, after staggering Hearns with an overhand right, Leonard pinned Hearns against the ropes, where he unleashed another furious combination, prompting referee Davey Pearl to stop the contest and award Sugar Ray Leonard the Unified World Welterweight Championship. Hearns was leading by scores of 124–122, 125–122, and 125–121.
After the fight, there was controversy due to the scoring of rounds six and seven. Even though Leonard dominated, hurting Hearns and battering him, all three judges gave both rounds to Leonard by a 10–9 margin. Many felt that the ten-point must scoring system was not properly used and those rounds should have been scored 10–8. Some also considered the stoppage premature. Veteran ringside commentator Don Dunphy said "They're stopping the fight. I don't believe it. Hearns was ahead on points." However, Emanuel Steward, Hearns' manager and trainer, said, "I felt that the referee was justified in stopping the fight ... Tommy did not have enough energy to make it through the fight."
The fight was named "Fight of the Year" by The Ring. Leonard was named "Fighter of the Year" by The Ring and The Boxing Writers Association of America. He was also named "Athlete of the Year" by ABC's Wide World of Sports and "Sportsman of the Year" by Sports Illustrated.
Retirement and return
On February 15, 1982, Leonard defended the unified title against Bruce Finch, the WBC fourth-ranked contender, in a bout at Reno, NV. Leonard knocked him out in the third round. Leonard's next fight was scheduled to be against Roger Stafford on May 14, 1982, in Buffalo, New York. While training, Leonard started to see floaters. He went to a doctor and discovered that he had a detached retina. The fight was cancelled, and Leonard had surgery to repair the retina on May 9, 1982.
On November 9, 1982, Leonard invited Marvin Hagler and other boxing dignitaries to a charity event in Baltimore, Maryland to hear him announce whether he would continue his career. Standing in a boxing ring with Howard Cosell, the master of ceremonies, Leonard announced his retirement, saying a bout with Hagler would unfortunately never happen. Leonard maintained his eye was fully healed, but that he just didn't want to box anymore.
Missing the limelight and the competition, Leonard announced in December 1983 that he was returning to the ring. Leonard boasted that he would have a couple of ten-round bouts and then take on Milton McCrory, Donald Curry, Durán, Hearns and finally Hagler. This decision was met with a torrent of criticism from fans and the media, who felt Leonard was taking unnecessary risks with his surgically repaired eye.
A bout with Philadelphia's Kevin Howard, who was 20–4–1, was scheduled for February 25, 1984. The fight was postponed when Leonard had minor surgery on his right eye to fix a loose retina. This latest eye problem further fueled the flames of those who opposed Leonard's comeback.
Before the fight with Howard, Dave Jacobs rejoined Leonard's team in a limited role. Jacobs had quit in 1980, disagreeing with Leonard's decision to have an immediate rematch with Durán.
Leonard and Howard fought on May 11, 1984, in Worcester, Massachusetts. Howard knocked Leonard flat on his back in the fourth round. It was the first knockdown of Leonard's professional career. Leonard came back to stop Howard in the ninth round, but the stoppage was disputed, with some feeling that the referee stopped the fight prematurely. Leonard was ahead on all three scorecards at the time of the stoppage. At the post-fight press conference, Leonard surprised everyone by announcing his retirement again, saying he just didn't have it anymore.
Leonard vs. Hagler
On March 10, 1986, Marvin Hagler knocked out John Mugabi in eleven rounds to retain the Undisputed World Middleweight Championship for the twelfth time and advance his record to 62–2–2. After the bout Hagler stated it may be his last fight. "I was ringside", Leonard said. "I'm watching John 'The Beast' Mugabi outbox Hagler. Of all people, John 'The Beast' Mugabi." It was then that Leonard decided to come back and fight Hagler. He called Mike Trainer and said, "I can beat Hagler".
On May 1, 1986, Leonard announced on a Washington, D.C. talk show that he would return to the ring to fight Hagler. The announcement generated a lot of controversy because of Leonard's inactivity and eye injuries, yet it also excited many sports fans who had hoped to see them fight years earlier. Hagler took a few months to decide, then agreed to the match.
The fight, promoted as "The Super Fight" and "The King of the Ring", was scheduled for April 6, 1987, at Caesars Palace in Las Vegas. Leonard was guaranteed $11 million, and Hagler was guaranteed $12 million. Hagler was a heavy favorite. The odds started at 4–1, then settled at 3–1. A paying crowd of 12,379 generated a live gate of $6.2 million. According to Bob Arum, the fight grossed $78 million (which equates to around $179 million in 2020).
The original fight plan for Leonard was to go toe-to-toe with Hagler and try to cut him, but the plan changed about five days before the fight. Leonard got hit by sparring partner Quincy Taylor and was badly buckled. "He almost knocked me out", Leonard said. After that, Leonard decided to box Hagler.
Many were surprised that Hagler, a natural southpaw, opened the fight boxing out of an orthodox stance. After the quick and slick Leonard won the first two rounds on all three scorecards, Hagler started the third round as a southpaw. Hagler did better, but Leonard's superior speed and boxing skill still allowed him to control the fight. Hagler looked stiff and mechanical and missed the speedy Leonard time and again prompting CBS ringside commentator Gil Clancy to remark "...and is he ever missing...Leonard isn't doing anything to make him miss, he's just missing!"
By the fifth, Leonard, who was moving a lot, began to tire and Hagler started to get closer. Hagler buckled Leonard's knees with a right uppercut near the end of the round, which finished with Leonard on the ropes. Hagler continued to score somewhat effectively in round six. Leonard, having slowed down, was obliged to fight more and move less. However, he was able to outpunch Hagler along the ropes and got the better of several bristling exchanges. Hagler never seized total control of the fight as he had against Thomas Hearns two years earlier, when he brutalized Hearns and scored a third-round knockout. Hagler's punches lacked snap and, although he was scoring solidly to the body, he looked nothing like the powerful fighter who had dominated the middleweight division for the previous five years. Leonard's observation that the Hagler who beat John Mugabi was older and slower proved to be spot on. In rounds seven and eight, Hagler's southpaw jab was landing solidly and Leonard's counter flurries were less frequent.
Round nine was the most exciting round of the fight. Hagler hurt Leonard with a left cross and pinned him in a corner. Leonard looked to be in trouble, but he furiously fought his way out of the corner. The action see-sawed back and forth for the rest of the round, with each man having his moments. However, Hagler's moments were more spectacular and one of Hagler's cornermen: Roger Perron (in an interview that took place on an episode of HBO's Legendary Nights episode segments in 2003) later stated that: "the ninth round was probably Marvin (Hagler)'s, best round".
Round ten was tame by comparison, as the pace slowed after the furious action of the previous round but with Hagler having more spectacular moments. Despite Leonard's obvious fatigue, he boxed well in the eleventh. Every time Hagler scored, Leonard came back with something flashier and more eye-catching, if not as effective. But at that point in the fight, Hagler appeared to be slightly more ring-general and clearly more aggressive. Between rounds eleven and twelve, Leonard's trainer: Angelo Dundee, implored Sugar Ray to get up off his stool yelling "We got three minutes...new champ...new champ!" Leonard yelled "Yeah!" and played to the screaming crowd. Hagler's corner was much more reserved prompting Clancy to comment: "They're talking to him like it's an IBM meeting or something...no emotion." In the final round, Hagler continued to chase Leonard. He hit Leonard with a big left hand and backed him into a corner. Leonard responded with a furious flurry, landing few punches but whipping the upset-hoping crowd into a frenzy. Hagler backed off, and Leonard danced away with Hagler in pursuit. The fight ended with Hagler and Leonard exchanging along the ropes. At the final bell, even uniformed ringside security rushed into the ring applauding and lauding Leonard's effort.
Leonard threw 629 punches and landed 306, while Hagler threw 792 and landed 291.
Leonard was awarded a controversial split-decision. Judge Dave Moretti scored it 115–113 for Leonard, while judge Lou Filippo had it 115–113 for Hagler. Judge José Guerra scored the fight 118–110 for Leonard. Many felt that Hagler deserved the decision because he was the aggressor and landed the harder punches. Scottish boxing journalist Hugh McIlvanney wrote that Leonard's plan was to "steal rounds with a few flashy and carefully timed flurries...he was happy to exaggerate hand speed at the expense of power, and neither he nor two of the scorers seemed bothered by the fact that many of the punches landed on the champion's gloves and arms."
Many others felt that Leonard deservedly got the decision, arguing that Leonard landed more punches and showed better defense and ring generalship. Jim Murray, long-time sports columnist for the Los Angeles Times, wrote, "It wasn't even close...He didn't just outpoint Hagler, he exposed him. He made him look like a guy chasing a bus. In snowshoes...Leonard repeatedly beat Hagler to the punch. When he did, he hit harder. He hit more often...He made Hagler into what he perceived him to be throughout his career—a brawler, a swarmer, a man who could club you to death only if you stood there and let him. If you moved, he was lost."
The scorecards from the ringside press and broadcast media attest to the polarizing views and opinions of the fight.
The fight was named "Fight of the Year" and "Upset of the Year" by The Ring.
Despite requests from the Hagler camp, Leonard was uninterested in a rematch and retired on May 27, 1987. "I'll try, I'll give it a shot", Leonard said of his latest retirement. "But you guys know me." A month after Hagler's formal retirement in June 1988, Leonard would announce another comeback.
Second comeback
Leonard vs. Lalonde
On November 7, 1988, Leonard made another comeback, facing Donny Lalonde at Caesar's Palace in Las Vegas. They fought for Lalonde's WBC Light Heavyweight Championship and the newly created WBC Super Middleweight Championship, which meant that Lalonde had to make 168 lbs. Many were critical of the fact that Lalonde's light heavyweight title was on the line when the weight limit of the fight with Leonard was at 168 pounds, and critical of Leonard for stipulating that his opponent—a natural 175-pounder—should weigh less than his usual fighting weight, which could possibly weaken him. However, Lalonde later told HBO's Larry Merchant that he didn't have any trouble making weight.
Lalonde, 31–2 with 26 knockouts, was guaranteed at least $6 million and Leonard was guaranteed over $10 million.
This would be Leonard's first professional fight without Angelo Dundee. For Leonard's fight with Hagler, Dundee worked without a contract and received $175,000, which was less than 2% of Leonard's purse. Dundee was unhappy with that amount. He requested a contract for the Lalonde fight and Leonard refused. "I don't have contracts. My word is my bond", Leonard said. Janks Morton and Dave Jacobs trained Leonard for the Lalonde fight.
Lalonde's size and awkwardness troubled Leonard. In the fourth round, a right hand to the top of Leonard's head dropped him for just the second time in his career. Early in the ninth, Lalonde hurt Leonard with a right to the chin. Leonard fired back and hurt Lalonde with a right. He drove him to the ropes and unleashed a furious assault. Lalonde tried to tie up Leonard, but got dropped with a powerful left hook. He rose but was soon down again, and the fight was stopped. Judges Chuck Giampa and Franz Marti had Leonard ahead by scores of 77–74 and 77–75, respectively. Judge Stuart Kirshenbaum had Lalonde ahead 76–75.
After the fight, Leonard vacated the light heavyweight title, but kept the super middleweight title. Also, Leonard and Janks Morton split because of personal differences. Morton was replaced as co-trainer by Pepe Correa, who had worked with Leonard for most of the previous fifteen years.
Leonard vs. Hearns
On June 12, 1989, Leonard defended the WBC Super Middleweight Championship in a rematch with Thomas Hearns at Caesar's Palace. It was promoted as "The War". Hearns was guaranteed $11 million and Leonard was guaranteed $14 million.
Hearns dropped Leonard with a right cross in the third round, but Leonard came back and battered Hearns around the ring in the fifth. Early in the seventh round, Hearns hurt Leonard but punched himself out going for the knockout. With Hearns fatigued, Leonard came back and had a strong finish to the round. Rounds nine and ten were good rounds for Leonard, but he ran into trouble in the eleventh round. Three booming rights from Hearns sent Leonard down for the second time in the fight. Knowing he needed a big finish, Leonard fought furiously and had a big final round.
The judges scored the fight a draw and Leonard retained the title. Judge Jerry Roth scored the fight 113–112 for Hearns, Judge Tom Kazmarek scored it 113–112 for Leonard, and Judge Dalby Shirley scored it 112–112. Shirley was the only judge to give Leonard a 10–8 margin in the twelfth. If he had scored it 10–9, as his two colleagues did, Hearns would have won by a split decision. Eventually, Leonard admitted that Hearns deserved the decision.
Leonard vs. Durán III – Uno Más
On December 7, 1989, Leonard defended the title against Roberto Durán, who was the reigning WBC Middleweight Champion. Durán was guaranteed $7.6 million and Leonard's arrangement guaranteed him over $13 million.
For the Durán fight, Leonard cut his entourage from twenty-one to six. Dave Jacobs was one of the people let go, leaving Correa as the sole trainer. Correa was instructed not to spare the whip. "For the first time in a long time, I allowed someone to push me", Leonard said.
The fight took place at the new Mirage Hotel in Las Vegas. Leonard used constant lateral movement and won by a lopsided twelve-round unanimous decision over a listless Durán. The scores were 120–110, 119–109, and 116–111. In a fight that many considered to be very boring, both fighters were booed often by the fans and many left the arena before the decision was announced. Pat Putnam of Sports Illustrated wrote, "Leonard gave them artistic perfection when they wanted heated battle, and they booed lustily. Most fight fans would not spend a dime to watch Van Gogh paint Sunflowers, but they would fill Yankee Stadium to see him cut off his ear." Although Leonard dominated the fight, he suffered several cuts. His lower lip was cut from a headbutt in the fourth round, his left eye was cut in the eleventh round, and his right eye was cut in the twelfth round. The cuts required a total of 60 stitches.
In August 1990, Leonard relinquished the WBC super middleweight title, saying that he was under the weight for the division. He then offered Hearns a third fight, but Hearns said he could no longer make the weight and moved up to the light heavyweight division.
Leonard vs. Norris
On February 9, 1991, Leonard went down to 154 lbs and fought WBC Light Middleweight Champion Terry Norris at Madison Square Garden. Leonard entered the bout as a 3-1 favorite but Norris dominated the fight, giving Leonard a heavy beating. He knocked Leonard down with a left hook in the second round, and in the seventh, he dropped Leonard again with a short right. Leonard had no answer for the skillful, younger, faster man. Leonard went the distance but lost by a lopsided decision. The scores were 120–104, 119–103, and 116–110. After the verdict was announced, Leonard announced his retirement. "It took this fight to show me it is no longer my time", Leonard said. "Tonight was my last fight. I know how Hagler felt now."
Final comeback
In October 1996, the 40-year-old Leonard announced that he was coming out of retirement to fight 34-year-old Héctor Camacho for the lightly regarded International Boxing Council (IBC) Middleweight Championship. Camacho, a light-hitting southpaw, was a three-time world champion with a record of 62–3–1. However, Camacho was also considered to be past his prime. Leonard decided to fight Camacho after commentating on Camacho's fight with the 45-year-old Roberto Durán the previous year, describing the disputed unanimous decision as "an early Christmas gift".
Leonard blamed his poor performance against Norris on lack of motivation, a rib injury, moving down in weight, and divorce, which was being litigated while he was in training. "It was stupid for me to fight Norris at 154 lbs", Leonard said. "This is different. I'm in the best shape possible."
For the Camacho fight, Leonard had a new trainer, Adrian Davis. "He's a great trainer, a throwback", Leonard said. "He has really helped me get ready."
In January 1997, it was announced that Leonard had been voted into the International Boxing Hall Of Fame in Canastota, New York. The rules state that a boxer must be retired for five years before being eligible for induction. When the vote took place, Leonard had been retired for more than five years, therefore, he was eligible, even though he had a fight scheduled. The induction ceremony was on June 15, 1997.
The fight with Camacho took place on March 1, 1997, in Atlantic City, New Jersey. Camacho applied pressure from the opening bell and started to score well in the third. He continued to score well in the fourth and opened a cut above Leonard's right eye. In the fifth, Camacho dropped Leonard with a right followed by two left uppercuts. Leonard got up, but was unable to ward off Camacho. The referee stopped the fight with Camacho teeing off on a defenseless Leonard on the ropes. It was the only time in Leonard's career that he was knocked out.
Afterward, Leonard retired again, saying, "For sure, my career is definitely over for me in the ring." However, less than a week after the fight, Leonard said he planned to fight again. He blamed his loss on a torn right calf muscle. His doctor suggested that he cancel the fight, but Leonard wanted to go through with it. Before the fight, he was given a shot of novocaine.
Leonard said he planned to have a series of tuneup fights before fighting a champion. He was scheduled to fight Tony Menefee on February 15, 1998, in Australia, but he pulled out of the fight, saying that he didn't have the motivation. The Camacho fight was Leonard's last. He finished his career with a record of 36–3–1 with 25 knockouts.
Media appearances
Leonard has worked as a boxing analyst for ABC, CBS, NBC, ESPN, HBO and EPIX. His relationship with HBO lasted for more than a decade. It ended in 1990, after HBO was not offered an opportunity to bid on the telecast rights to Leonard's fight with Terry Norris. HBO believed it would be inappropriate for Leonard to continue with them if they couldn't bid on his fights. Leonard's attorney, Mike Trainer, said, "There never has been a linkage between his broadcasting and his fighting."
Leonard has provided commercial endorsements for companies including Coca-Cola, EA Sports, Ford, Nabisco, Revlon and 7 Up. His most famous commercial was a 7 Up ad he did with his son, Ray Jr., Roberto Durán and Durán's son Roberto Jr. in the early 1980s. Leonard is among the most sought-after motivational/inspirational speakers in the world today. His speech, entitled "Power" (Prepare, Overcome and Win Every Round), is consistently booked with major Fortune 500 companies throughout the United States and abroad.
Leonard has also worked as an actor. He has appeared in numerous television shows, including Half & Half, L.A. Heat, Married... with Children, Renegade and Tales From The Crypt. He has also appeared in several movies, including I Spy and most recently The Fighter (2010), starring Christian Bale and Mark Wahlberg. This movie brought back memories of his fight with Dicky Eklund. He also worked as an adviser in the 2011 robot boxing film Real Steel. Leonard served as host and mentor to the aspiring fighters on The Contender. Sylvester Stallone, who co-hosted during the first season, was one of the executive producers, along with Mark Burnett. When Leonard left the show, he was replaced as host by Tony Danza for the final season.
In 2001, Leonard launched Sugar Ray Leonard Boxing Inc., a boxing promotional company, and announced the company's strategic partnership with ESPN. Together, Leonard and ESPN would produce and promote "Sugar Ray Leonard and ESPN II Presents Friday Night Fights", which would air the first Friday of every month for twelve months. Leonard's boxing promotional company was dissolved in 2004. He had a falling out with partner Bjorn Rebney, whom he called "a cancer in my company." Speaking of his promotional company, Leonard said, "We did some great shows with evenly matched fights. I took great pride in it. But the TV show came about and made my decision a lot easier. I already had it in the back of my mind to dissolve the company. The working environment was not healthy."
Leonard competed on season 12 of Dancing with the Stars, which premiered on Monday, March 21, 2011, on ABC. His partner was Anna Trebunskaya. He was voted off in Week 4 of the show. During his appearance on The Colbert Report in 2011, Leonard was defeated by host Stephen Colbert in a thumb wrestling contest. He appeared as a guest at the chef's table, along with Tito Ortiz, during the tenth season of Hell's Kitchen. He is the celebrity spokesperson for the Atlanta law firm John Foy and Associates, PC.
Leonard was also the subject of a Seinfeld episode (season 6, episode 22) where George tries to flatter his boss by saying he looks like Sugar Ray Leonard. The real Leonard (a Seinfeld fan) mentioned that he was told about the episode by friends and family, but had never seen it for himself until a friend gave him the DVD set for a gift.
In March 2019, Leonard made a guest appearance on the popular daily morning show on YouTube, Good Mythical Morning, hosted by Rhett McLaughlin and Link Neal. In this episode of the "March Milkness" series (where Rhett and Link created a March Madness style bracket to determine the best breakfast cereal), Leonard was brought in to break a tie between the Fruity Region Champion, Froot Loops, and the Chocolatey Region Champion, Oreo O's. His decision was to push Froot Loops onto the Finals, where they were taken out by Cinnamon Toast Crunch.
Personal life
Family
Leonard married his high school sweetheart, Juanita Wilkinson, in January 1980. Their six-year-old son, Ray Jr., served as the ring bearer. In 1984, they had another son, Jarrell. They were divorced in 1990. During divorce proceedings, Juanita Leonard testified that her husband physically abused her while under the influence of alcohol. She also said he was an occasional cocaine user. In his testimony, Leonard confirmed his wife's claims and went on to reveal that the problems of their marriage were not due to drug and alcohol use.
After the Los Angeles Times broke the story, Leonard held a press conference and publicly acknowledged that the accusations were true. He said he started using medication after he retired in 1982, following surgery to repair a detached retina. "I wanted more", Leonard said. "I wanted that arena. I didn't want anyone to tell me my career had to end." "I decided to search for a substitute...I resorted to cocaine. I used when I felt bad, I used when I missed competing at that level", he said. "It was a crutch, something that enabled me to forget." He said he quit using drugs in early 1986, when he woke up one morning and "what I saw in the mirror was scary." "I can never erase the pain or the scars I have made through my stupidity, my selfishness", Leonard said. "All I can do is say I'm sorry, but that is not enough." In 2011, Leonard revealed in an NPR interview that he had been free of alcohol since July 2006.
In 1989, Leonard was introduced to Bernadette Robi by Kenny G at a Luther Vandross concert. Robi is the daughter of Paul Robi, one of the original Platters, and she is the ex-wife of Lynn Swann. Leonard and Robi were married at Leonard's $8.7 million estate in Pacific Palisades, California in August 1993. At the wedding ceremony, the grounds were converted into a garden with 10,000 roses and blossoms of other flowers flown in from the Netherlands.
Leonard is also the godfather of Khloé Kardashian and has appeared on many episodes of Keeping Up With The Kardashians.
Charity work
For many years, Leonard has been the International Chairman of the Juvenile Diabetes Research Foundation's Walk for a Cure and is actively involved in raising both awareness and funds.
Leonard testified before the Senate Committee on Homeland Security and Government Affairs in 2009. The Senate hearing was titled "Type 1 Diabetes Research: Real Progress and Real Hope for a Cure". He testified about the burden of diabetes and the need for continued research funding to find a cure.
Leonard and his wife, Bernadette, founded the Sugar Ray Leonard Foundation to support the Juvenile Diabetes Research Foundation and its annual Walk for a Cure. In 2009, the foundation expanded to support programs that help people rebuild their communities in ten cities across the United States. It supports accessible housing, healthcare services, and educational services and job training.
In 2007 he was awarded The Ambassador Award of Excellence by the LA Sports & Entertainment Commission at the Riviera Country Club for his continued community involvement.
Advocacy against child molestation
In his autobiography The Big Fight: My Life in and out of the Ring, published in June 2011, Leonard reveals that as a young boxer he was the victim of sexual abuse from an Olympic trainer as well as another man, a benefactor. He has since made public appearances to bring attention to the issue of child sex abuse, declaring himself a "poster child" for the cause and encouraging victims to report their abuse.
Professional boxing record
See also
List of welterweight boxing champions
List of light middleweight boxing champions
List of middleweight boxing champions
List of undisputed boxing champions
List of WBA world champions
List of WBC world champions
List of WBO world champions
List of The Ring world champions
List of boxing quintuple champions
Sugar Ray Leonard vs. Roberto Durán II
Sugar Ray Leonard vs. Thomas Hearns
References
External links
1956 births
Living people
African-American boxers
African-American male actors
American male actors
Boxers at the 1976 Summer Olympics
American boxing promoters
International Boxing Hall of Fame inductees
Light-middleweight boxers
Light-heavyweight boxers
Middleweight boxers
Boxers at the 1975 Pan American Games
Pan American Games gold medalists for the United States
Olympic boxers of the United States
Olympic gold medalists for the United States in boxing
Medalists at the 1976 Summer Olympics
Participants in American reality television series
Super-middleweight boxers
World Boxing Association champions
World Boxing Council champions
World Boxing Organization champions
Welterweight boxers
Winners of the United States Championship for amateur boxers
Boxers from Washington, D.C.
Sportspeople from Wilmington, North Carolina
People from Palmer Park, Maryland
American male boxers
The Ring (magazine) champions
World welterweight boxing champions
World light-middleweight boxing champions
World middleweight boxing champions
World super-middleweight boxing champions
World light-heavyweight boxing champions
Pan American Games medalists in boxing
Light-welterweight boxers
Lightweight boxers
National Golden Gloves champions
Boxers from North Carolina
Sportspeople with a vision impairment
Medalists at the 1975 Pan American Games
Boxing commentators
21st-century African-American people
20th-century African-American sportspeople |
null | null | Menstrual cycle | eng_Latn | The menstrual cycle is a series of natural changes in hormone production and the structures of the uterus and ovaries of the female reproductive system that make pregnancy possible. The ovarian cycle controls the production and release of eggs and the cyclic release of estrogen and progesterone. The uterine cycle governs the preparation and maintenance of the lining of the uterus (womb) to receive a fertilized egg. These cycles are concurrent and coordinated, normally last between 21 and 35 days in adult women, with a median length of 28 days, and continue for about 30–45 years.
Naturally occurring hormones drive the cycles; the cyclical rise and fall of the follicle stimulating hormone prompts the production and growth of oocytes (immature egg cells). The hormone estrogen stimulates the uterus lining to thicken to accommodate an embryo should fertilization occur. The blood supply of the thickened lining (endometrium) provides nutrients to a successfully implanted embryo. If implantation does not occur, the lining breaks down and blood is released. Triggered by falling progesterone levels, menstruation (a "period", in common parlance) is the cyclical shedding of the lining, and is a sign that pregnancy has not occurred.
Each cycle occurs in phases based on events in the ovary (ovarian cycle) or the uterus (uterine cycle). The ovarian cycle consists of the follicular phase, ovulation, and the luteal phase; the uterine cycle consists of the menstrual, proliferative and secretory phases. Day one of the menstrual cycle is the first day of the period, which lasts for about five days. Around day fourteen, an egg is usually released from the ovary. Menarche (the onset of the first period) usually occurs around the age of twelve years.
The menstrual cycle can cause some women to experience problems that disrupt daily lives. These can include cramps, tender breasts, tiredness, and premenstrual syndrome. More severe problems such as premenstrual dysphoric disorder are experienced by 3–8% of women. The menstrual cycle can be modified by hormonal birth control.
Cycles and phases
The menstrual cycle encompasses the ovarian and uterine cycles. The ovarian cycle describes changes that occur in the follicles of the ovary, whereas the uterine cycle describes changes in the endometrial lining of the uterus. Both cycles can be divided into phases. The ovarian cycle consists of alternating follicular and luteal phases, and the uterine cycle consists of menstruation, the proliferative phase, and the secretory phase. The menstrual cycle is controlled by the hypothalamus and the pituitary gland in the brain. The hypothalamus releases gonadotropin-releasing hormone (GnRH), which causes the nearby anterior pituitary gland to release follicle-stimulating hormone (FSH) and luteinizing hormone (LH). Before puberty, GnRH is released in low steady quantities and at a steady rate. After puberty, GnRH is released in large pulses, and the frequency and magnitude of these determine how much FSH and LH are produced by the pituitary.
Measured from the first day of one menstruation to the first day of the next, the length of a menstrual cycle varies but has a median length of 28 days. The cycle is often less regular at the beginning and end of a woman's reproductive life. At puberty, a child's body begins to mature into an adult body capable of sexual reproduction; the first period (called menarche) occurs at around 12 years of age and continues for about 30–45 years. Menstrual cycles end at menopause, which is usually between 45 and 55 years of age.
Ovarian cycle
Between menarche and menopause the human ovaries regularly alternate between luteal and follicular phases during the monthly menstrual cycle. Stimulated by gradually increasing amounts of estrogen in the follicular phase, discharges of blood flow stop and the uterine lining thickens. Follicles in the ovary begin developing under the influence of a complex interplay of hormones, and after several days one, or occasionally two, become dominant, while non-dominant follicles shrink and die. About mid-cycle, some 10–12 hours after the luteinizing hormone (LH) surges, the dominant follicle releases an oocyte, in an event called ovulation.
After ovulation, the oocyte lives for 24 hours or less without fertilization, while the remains of the dominant follicle in the ovary become a corpus luteum – a body with the primary function of producing large amounts of the hormone progesterone. Under the influence of progesterone, the uterine lining changes to prepare for potential implantation of an embryo to establish a pregnancy. The thickness of the endometrium continues to increase in response to mounting levels of estrogen, which is released by the antral follicle (a mature ovarian follicle) into the blood circulation. Peak levels of estrogen are reached at around day thirteen of the cycle and coincide with ovulation. If implantation does not occur within about two weeks, the corpus luteum degenerates into the corpus albicans, which does not produce hormones, causing a sharp drop in levels of both progesterone and estrogen. This drop causes the uterus to lose its lining in menstruation; it is around this time that the lowest levels of estrogen are reached.
In an ovulatory menstrual cycle, the ovarian and uterine cycles are concurrent and coordinated and last between 21 and 35 days in an adult woman, with a population average of 27–29 days. Although the average length of the human menstrual cycle is similar to that of the lunar cycle, there is no causal relation between the two.
Follicular phase
The ovaries contain a finite number of egg stem cells, granulosa cells and theca cells, which together form primordial follicles. At around 20 weeks into gestation some 7 million immature eggs have already formed in an ovary. This decreases to around 2 million by the time a girl is born, and 300,000 by the time she has her first period. On average, one egg matures and is released during ovulation each month after menarche. Beginning at puberty, these mature to primary follicles independently of the menstrual cycle. The development of the egg is called oogenesis and only one cell survives the divisions to await fertilization. The other cells are discarded as polar bodies, which cannot be fertilized. The follicular phase is the first part of the ovarian cycle and it ends with the completion of the antral follicles. Meiosis (cell division) remains incomplete in the egg cells until the antral follicle is formed. During this phase usually only one ovarian follicle fully matures and gets ready to release an egg. The follicular phase shortens significantly with age, lasting around 14 days in women aged 18–24 compared with 10 days in women aged 40–44.
Through the influence of a rise in follicle stimulating hormone (FSH) during the first days of the cycle, a few ovarian follicles are stimulated. These follicles, which have been developing for the better part of a year in a process known as folliculogenesis, compete with each other for dominance. All but one of these follicles will stop growing, while one dominant follicle – the one that has the most FSH receptors – will continue to maturity. The remaining follicles die in a process called follicular atresia. Luteinizing hormone (LH) stimulates further development of the ovarian follicle. The follicle that reaches maturity is called an antral follicle, and it contains the ovum (egg cell).
The theca cells develop receptors that bind LH, and in response secrete large amounts of androstenedione. At the same time the granulosa cells surrounding the maturing follicle develop receptors that bind FSH, and in response start secreting androstenedione, which is converted to estrogen by the enzyme aromatase. The estrogen inhibits further production of FSH and LH by the pituitary gland. This negative feedback regulates levels of FSH and LH. The dominant follicle continues to secrete estrogen, and the rising estrogen levels make the pituitary more responsive to GnRH from the hypothalamus. As estrogen increases this becomes a positive feedback signal, which makes the pituitary secrete more FSH and LH. This surge of FSH and LH usually occurs one to two days before ovulation and is responsible for stimulating the rupture of the antral follicle and release of the oocyte.
Ovulation
Around day fourteen, the egg is released from the ovary. Called "ovulation", this occurs when a mature egg is released from the ovarian follicles into the fallopian tube, about 10–12 hours after the peak in LH surge. Typically only one of the 15–20 stimulated follicles reaches full maturity, and just one egg is released. Ovulation only occurs in around 10% of cycles during the first two years following menarche, and by the age of 40–50, the number of ovarian follicles is depleted. LH initiates ovulation at around day 14 and stimulates the formation of the corpus luteum. Following further stimulation by LH, the corpus luteum produces and releases estrogen, progesterone, relaxin (which relaxes the uterus by inhibiting contractions of the myometrium), and inhibin (which inhibits further secretion of LH).
The release of LH matures the egg and weakens the follicle wall in the ovary, causing the fully developed follicle to release its oocyte. If it is fertilized by a sperm, the oocyte promptly matures into an ootid, which blocks the other sperm cells and becomes a mature egg. If it is not fertilized by a sperm, the oocyte degenerates. The mature egg has a diameter of about , and is the largest human cell.
Which of the two ovaries – left or right – ovulates appears random; no left and right coordinating process is known. Occasionally both ovaries release an egg; if both eggs are fertilized, the result is fraternal twins. After release from the ovary, the egg is swept into the fallopian tube by the fimbria – a fringe of tissue at the end of each fallopian tube. After about a day, an unfertilized egg disintegrates or dissolves in the fallopian tube, and a fertilized egg reaches the uterus in three to five days.
Fertilization usually takes place in the ampulla, the widest section of the fallopian tubes. A fertilized egg immediately starts the process of embryogenesis (development). The developing embryo takes about three days to reach the uterus, and another three days to implant into the endometrium. It has usually reached the blastocyst stage at the time of implantation: this is when pregnancy begins. The loss of the corpus luteum is prevented by fertilization of the egg. The syncytiotrophoblast (the outer layer of the resulting embryo-containing blastocyst that later becomes the outer layer of the placenta) produces human chorionic gonadotropin (hCG), which is very similar to LH and preserves the corpus luteum. During the first few months of pregnancy, the corpus luteum continues to secrete progesterone and estrogens at slightly higher levels than those at ovulation. After this and for the rest of the pregnancy, the placenta secretes high levels of these hormones – along with human chorionic gonadotropin (hCG), which stimulates the corpus luteum to secrete more progesterone and estrogens, blocking the menstrual cycle. These hormones also prepare the mammary glands for milk production.
Luteal phase
Lasting about 14 days, the luteal phase is the final phase of the ovarian cycle and it corresponds to the secretory phase of the uterine cycle. During the luteal phase, the pituitary hormones FSH and LH cause the remaining parts of the dominant follicle to transform into the corpus luteum, which produces progesterone. The increased progesterone starts to induce the production of estrogen. The hormones produced by the corpus luteum also suppress production of the FSH and LH that the corpus luteum needs to maintain itself. The level of FSH and LH fall quickly, and the corpus luteum atrophies. Falling levels of progesterone trigger menstruation and the beginning of the next cycle. From the time of ovulation until progesterone withdrawal has caused menstruation to begin, the process typically takes about two weeks. For an individual woman, the follicular phase often varies in length from cycle to cycle; by contrast, the length of her luteal phase will be fairly consistent from cycle to cycle at 10 to 16 days (average 14 days).
Uterine cycle
The uterine cycle has three phases: menses, proliferative and secretory.
Menstruation
Menstruation (also called menstrual bleeding, menses or a period) is the first and most evident phase of the uterine cycle and first occurs at puberty. Called menarche, the first period occurs at the age of around twelve or thirteen years. The average age is generally later in the developing world and earlier in developed world. In precocious puberty, it can occur as early as age eight years, and this can still be normal.
Menstruation is initiated each month by falling levels of estrogen and progesterone and the release of prostaglandins, which constrict the spiral arteries. This causes them to spasm, contract and break up. The blood supply to the endometrium is cut off and the cells of the top layer of the endometrium (the stratum functionalis) become deprived of oxygen and die. Later the whole layer is lost and only the bottom layer, the stratum basalis, is left in place. An enzyme called plasmin breaks up the blood clots in the menstrual fluid, which eases the flow of blood and broken down lining from the uterus. The flow of blood continues for 2–6 days and around 30–60 milliliters of blood is lost, and is a sign that pregnancy has not occurred.
The flow of blood normally serves as a sign that a woman has not become pregnant, but this cannot be taken as certainty, as several factors can cause bleeding during pregnancy. Menstruation occurs on average once a month from menarche to menopause, which corresponds with a woman's fertile years. The average age of menopause in women is 52 years, and it typically occurs between 45 and 55 years of age. Menopause is preceded by a stage of hormonal changes called perimenopause.
Eumenorrhea denotes normal, regular menstruation that lasts for around the first 5 days of the cycle. Women who experience menorrhagia (heavy menstrual bleeding) are more susceptible to iron deficiency than the average person.
Proliferative phase
The proliferative phase is the second phase of the uterine cycle when estrogen causes the lining of the uterus to grow and proliferate. The latter part of the follicular phase overlaps with the proliferative phase of the uterine cycle. As they mature, the ovarian follicles secrete increasing amounts of estradiol, an estrogen. The estrogens initiate the formation of a new layer of endometrium in the uterus with the spiral arterioles.
As estrogen levels increase, cells in the cervix produce a type of cervical mucus that has a higher pH and is less viscous than usual, rendering it more friendly to sperm. This increases the chances of fertilization, which occurs around day 11 to day 14. This cervical mucus can be detected as a vaginal discharge that is copious and resembles raw egg whites. For women who are practicing fertility awareness, it is a sign that ovulation may be about to take place, but it does not mean ovulation will definitely occur.
Secretory phase
The secretory phase is the final phase of the uterine cycle and it corresponds to the luteal phase of the ovarian cycle. During the secretory phase, the corpus luteum produces progesterone, which plays a vital role in making the endometrium receptive to the implantation of a blastocyst (a fertilized egg, which has begun to grow). Glycogen, lipids, and proteins are secreted into the uterus and the cervical mucus thickens. In early pregnancy progesterone also increases blood flow and reduces the contractility of the smooth muscle in the uterus and raises the woman's basal body temperature.
If pregnancy does not occur the ovarian and uterine cycles start over again.
Anovulatory cycles and short luteal phases
Only two thirds of overtly normal menstrual cycles are ovulatory, that is, cycles in which ovulation occurs. The other third lack ovulation or have a short luteal phase (less than ten days) in which progesterone production is insufficient for normal physiology and fertility. Cycles in which ovulation does not occur (anovulation) are common in girls who have just begun menstruating and in women around menopause. During the first two years following menarche, ovulation is absent in around half of cycles. Five years after menarche, ovulation occurs in around 75% of cycles and this reaches 80% in the following years. Anovulatory cycles are often overtly identical to normally ovulatory cycles. Any alteration to balance of hormones can lead to anovulation. Stress, anxiety and eating disorders can cause a fall in GnRH, and a disruption of the menstrual cycle. Chronic anovulation occurs in 6–15% of women during their reproductive years. Around menopause, hormone feedback dysregulation leads to anovulatory cycles. Although anovulation is not considered a disease, it can be a sign of an underlying condition such as polycystic ovary syndrome. Anovulatory cycles or short luteal phases are normal when women are under stress or athletes increasing the intensity of training. These changes are reversible as the stressors decrease or, in the case of the athlete, as she adapts to the training.
Menstrual health
Although a normal and natural process, some women experience problems sufficient to disrupt their lives as a result of their menstrual cycle. These include acne, tender breasts, feeling tired, and premenstrual syndrome (PMS). More severe problems such as premenstrual dysphoric disorder are experienced by 3 to 8% of women. Dysmenorrhea or "period pain" can cause cramps in the abdomen, back, or upper thighs that occur during the first few days of menstruation. Debilitating period pain is not normal and can be a sign of something severe such as endometriosis. These issues can significantly affect a woman's health and quality of life and timely interventions can improve the lives of these women.
There are common culturally communicated misbeliefs that the menstrual cycle affects women's moods, causes depression or irritability, or that menstruation is a painful, shameful or unclean experience. Often a woman's normal mood variation is falsely attributed to the menstrual cycle. Much of the research is weak, but there appears to be a very small increase in mood fluctuations during the luteal and menstrual phases, and a corresponding decrease during the rest of the cycle. Changing levels of estrogen and progesterone across the menstrual cycle exert systemic effects on aspects of physiology including the brain, metabolism, and musculoskeletal system. The result can be subtle physiological and observable changes to women's athletic performance including strength, aerobic, and anaerobic performance. Changes to the brain have also been observed throughout the menstrual cycle but do not translate into measurable changes in intellectual achievement – including academic performance, problem-solving, memory, and creativity. Improvements in spatial reasoning ability during the menstruation phase of the cycle are probably caused by decreases in levels of estrogen and progesterone.
In some women, ovulation features a characteristic pain called mittelschmerz (a German term meaning middle pain). The cause of the pain is associated with the ruptured follicle, causing a small amount of blood loss.
Even when normal, the changes in hormone levels during the menstrual cycle can increase the incidence of disorders such as autoimmune diseases, which might be caused by estrogen enhancement of the immune system.
Around 40% of women with epilepsy find that their seizures occur more frequently at certain phases of their menstrual cycle. This catamenial epilepsy may be due to a drop in progesterone if it occurs during the luteal phase or around menstruation, or a surge in estrogen if it occurs at ovulation. Women who have regular periods can take medication just before and during menstruation. Options include progesterone supplements, increasing the dose of their regular anticonvulsant drug, or temporarily adding an anticonvulsant such as clobazam or acetazolamide. If this is ineffective, or when a woman's menstrual cycle is irregular, then treatment is to stop the menstrual cycle occurring. This may be achieved using medroxyprogesterone, triptorelin or goserelin, or by sustained use of oral contraceptives.
Hormonal contraception
Hormonal contraceptives prevent pregnancy by inhibiting the secretion of the hormones, FSH, LH and GnRH. Hormonal contraception that contains estrogen, such as combined oral contraceptive pills (COCs, often referred to as birth control pills) stop the development of the dominant follicle and the mid-cycle LH surge and thus ovulation. Sequential dosing and discontinuation of the COC can mimic the uterine cycle and produce bleeding that resembles a period. In some cases, this bleeding is lighter.
Progestin-only methods of hormonal contraception do not always prevent ovulation but instead work by stopping the cervical mucus from becoming sperm-friendly. Hormonal contraception is available in a variety of forms such as pills, patches, skin implants and hormonal intrauterine devices (IUDs).
Evolution and other species
Most female mammals have an estrous cycle, but only ten primate species, four bat species, the elephant shrews and the spiny mouse species Cairo spiny mouse (Acomys cahirinus) have a menstrual cycle. The cycles are the same as in humans apart from the length, which ranges from 9 to 37 days. The lack of immediate relationship between these groups suggests that four distinct evolutionary events have caused menstruation to arise. In species that have a menstrual cycle, ovulation is not obvious to potential mates and there is no mating season. There are four theories on the evolutionary significance of menstruation:
Control of sperm-borne pathogens. This hypothesis held that menstruation protected the uterus against pathogens introduced by sperm. Hypothesis 1 does not take into account that copulation can take place weeks before menstruation and that potentially infectious semen is not controlled by menstruation in other species.
Energy conservation. This hypothesis claimed that it took less energy to rebuild a uterine lining than to maintain it if pregnancy did not occur. Hypothesis 2 does not explain other species that also do not maintain a uterine lining but do not menstruate.
A theory based on spontaneous decidualization (a process that results in significant changes to cells of the endometrium in preparation for, and during, pregnancy, in which the endometrium changes into the decidua). Decidualization leads to the development of the endothelium, which involves cells of the immune system, the formation of a new blood supply, hormones and tissue differentiation. In non-menstruating mammals, decidualization is driven by the embryo, not the mother. It evolved in some placental mammals because it confers advantages in that it allows females to prepare for pregnancy without needing a signal from the fetus. Hypothesis 3 defers to an explanation of the evolutionary origin of spontaneous decidualization and does not explain the evolution of menstruation alone.
Uterine pre-conditioning. This hypothesis claims that a monthly pre-conditioning of the uterus is needed in species, such as humans, that have deeply invasive (deep-rooted) placentas. In the process leading to the formation of a placenta, maternal tissues are invaded. This hypothesis holds that menstruation was not evolutionary, rather the result of a coincidental pre-conditioning of the uterus to protect uterine tissue from the deeply rooting placenta, in which a thicker endometrium develops. Hypothesis 4 does not explain menstruation in non-primates.
Notes
References
Book sources
External links
Menstrual cycle
Human female endocrine system
Midwifery
Periodic phenomena
Women's health
Human female reproductive system |
null | null | The Gong Show | eng_Latn | The Gong Show is an American amateur talent contest franchised by Sony Pictures Television to many countries. It was broadcast on NBC's daytime schedule from June 14, 1976, through July 21, 1978, and in first-run syndication from 1976 to 1980 and 1988 to 1989, and was revived in 2017 for broadcast on ABC. The show was created and originally produced by Chuck Barris, who also served as host for the NBC run and from 1977 to 1980 in syndication. Its most recent version was executive-produced by Will Arnett and hosted by Tommy Maitland, a fictional character performed by Mike Myers (uncredited in Season 1). The Gong Show is known for its absurdist humor and style, with the actual competition secondary to the often outlandish acts presented; a small cash prize has typically been awarded to each show's winner.
Format
Each show presented a competition of amateur performers of often dubious talent, with a panel of three celebrity judges. The original program's regular judges included Jamie Farr, Jaye P. Morgan, Arte Johnson, Patty Andrews, Phyllis Diller, Pat McCormick, Anson Williams, Steve Garvey, Rex Reed and Rip Taylor. Other celebrities occasionally appeared as judges. If any judge considered an act to be particularly bad, they could force it to stop by striking a large gong, a trope adapted from the durable radio show Major Bowes Amateur Hour. Barris would then ask the judge(s) in question why they had gonged the act, usually receiving a facetious response.
Any act that survived without being gonged was given a score by each of the three judges on a scale of 0 to 10, for a maximum possible score of 30. On the NBC series, the contestant who achieved the highest combined score won the grand prize: a check for $516.32 (a "highly unusual amount", in Barris's words; reportedly the Screen Actors Guild's minimum pay for a day's work at the time) and a "Golden Gong" trophy. The syndicated series' top prize was originally $712.05 (the first episode was $996.83) and later increased to $716.32. In the event of a tie, three different tiebreakers were used at various times during the show's run. Originally the studio audience determined the winner by applause, but this was later changed to a decision by the producers, and later by the celebrity judges. On rare occasions, both winning acts would each receive a check and a trophy. No prize was awarded if all of the acts on a particular episode were gonged, which occurred at least twice. Runners-up received various prizes; Maureen Orth, on her February 24, 1977, appearance, reported receiving an iron valued at $33.95 for her second-place finish.
When Barris announced the final score, actor Jerry Maren (a little person and former Munchkin) ran onstage in top hat and tails, throwing confetti while balloons dropped from overhead.
The daily Gong Show also gave out a "Worst Act of the Week" award (later changed to the "Most Outrageous Act of the Week"), selected by the producers and each week's judges. The winner of this award was announced following the trophy presentation on the Friday show, and the performer received a dirty tube sock and a check for $516.32.
Legitimate talent
The two biggest Gong Show-related showbiz successes were Andrea McArdle and Cheryl Lynn. Twelve-year-old McArdle appeared on an early show in 1976, shortly before winning the lead role in the hit Broadway musical Annie. Lynn was signed to a recording contract as a result of her performance, and recorded the Top 40 disco hit "Got To Be Real".
Among the other true talents that appeared on the show were country singer Boxcar Willie; comics and actors Paul Reubens and John Paragon (best known as Pee Wee Herman and Jambi the Genie); Joey D'Auria ("Professor Flamo", later WGN's second Bozo the Clown); impressionist/comic Michael Winslow; novelty rock band Green Jellÿ; and a band called The Mystic Knights of the Oingo Boingo which evolved into Oingo Boingo, led by future film and television score composer Danny Elfman. Crips founder and executed murderer Stanley Tookie Williams appeared on the show in 1979 as a bodybuilder. In 1976, future Academy Award-nominated actress Mare Winningham sang the Beatles song "Here, There, and Everywhere". Future Super Bowl XXXV winning head coach Brian Billick also made an appearance, performing a routine known as the "spider monkey". Dancer Danny Lockin, who had played Barnaby in the film Hello Dolly!, was murdered hours after winning the show taped August 21, 1977. Journalist Maureen Orth, then writing for Newsweek, appeared in 1977 with a second-place showing, appearing as "The World's Oldest Cheerleader."
Personnel
Barris as emcee
An established game show producer (The Dating Game, The Newlywed Game), Barris was originally the show's co-producer but not its host. He was an emergency replacement host for eventual Real People host John Barbour, who had objected to the show's satirical concept and tried to steer it toward a traditional amateur-hour format.
Producer Chris Bearde, formerly of Rowan and Martin's Laugh-In and The Sonny & Cher Comedy Hour, clashed with Barris over the show's content, favoring scripted comedy over chaotic nonsense. (Bearde's "new talent" segments on Laugh-In had featured oddball performers, the most famous being Tiny Tim.) Bearde eventually resigned from The Gong Show, leaving Barris fully in charge. Before long, Barris was working so loosely that some viewers assumed he was intoxicated from alcohol or other drugs. Barris later recounted, in an interview with the Archive of American Television, that he was never drunk, and that he would not allow the use of drugs in his production company.
Musical direction
Milton DeLugg, the show's musical director, was a popular musician and bandleader during the 1940s, and got the Gong Show job by default. As musical director for the network, he was responsible for any NBC project that required special music (like the annual telecasts of the Macy's Thanksgiving Day Parade). Barris initially regarded DeLugg as "an anachronism", but he soon found that DeLugg was very much attuned to the crazy tone of the show. He appeared in recurring comedy skits with Barris.
Veteran composer Joey Carbone provided musical arrangements for the late 1980s revival with his own lineup of studio musicians, known as "The Gong Show Guys".
Announcers
Johnny Jacobs, who had worked for Barris for many years, was the main announcer from 1976 to 1980. When Jacobs was sidelined with an extended illness, Jack Clark substituted from October 3 through December 23, 1977. Charlie O'Donnell served as announcer for the late 1980s revival.
Hostesses
Hostesses included Siv Åberg (a Swedish-born model and actress who appeared on Barris's syndicated New Treasure Hunt), actress Marlena Clark, porn star Carol Connors and Barris's teenage daughter, Della.
Broadcast history
NBC
NBC first aired the show at 12:30 p.m. (11:30 a.m. Central). This was the network's least important time slot, as programs running at that time had to share the half-hour with a five-minute NBC newscast anchored by Edwin Newman. As a result, the first six-plus months of The Gong Show featured approximately twenty minutes of program content in a twenty-five-minute episode.
Many NBC affiliates in some larger markets opted not to run network programming during the noon hour at all, preferring to broadcast local news and talk shows instead. Thus Gong made its debut mainly on medium-market and smaller stations or on large-market rival stations that had picked up the program from the NBC affiliate that had rejected it. For example, in Boston, then-NBC affiliate WBZ did not run the series, allowing local UHF independent outlet WSBK-TV to air it.
Gong's time slot was given to a new soap opera, Lovers and Friends, on January 3, 1977, and the show was relocated to replace the cancelled Another World spinoff Somerset at 4:00 p.m. The time change allowed Gong to expand to a half-hour.
NBC broadcast a one-hour prime-time Gong Show special on April 26, 1977, featuring in-studio special guests Tony Randall, Alice Cooper and Harry James and His Orchestra. The winning act on this special was The Bait Brothers, and the panelists were Jaye P. Morgan, Jamie Farr and Arte Johnson.
Popsicle Twins incident
During The Gong Show'''s run, Barris became well known for his clashes with the network censors, intentionally bringing in risque acts as bait to allow some of the less racy acts to slip by. In 1978, one of these bait acts, two teenage girls referring to themselves as "Have You Got a Nickel?", made it onto the show. Their act consisted of the girls sitting cross-legged on the stage floor and silently eating popsicles in a manner that suggested they were performing fellatio on the frozen treats. The nature of the act led to the two girls being known as the "Popsicle Twins".
While they were able to complete their act without being gonged, two of the judges gave them low marks. Phyllis Diller gave them a zero, while Jamie Farr awarded them a marginally better 2. Jaye P. Morgan awarded them a 10, quipping, "Do you know that that's the way I started?"
Surprisingly, the girls' act was approved by the censors, who apparently did not see anything wrong with it during the rehearsals. However, after the episode aired in the Eastern Time Zone, NBC cut the act from the later tape delay broadcast for western time zones. The act was not cut from all the tapes, and the Popsicle Twins incident has aired in reruns and retrospectives. Barris said in a 2001 interview with Salon.com that this particular act began making him reconsider his career.
Cancellation
Despite its popularity and respectable ratings for a non-soap-opera midday show, NBC cancelled The Gong Show, with its final episode broadcast on July 21, 1978. At the time, there was much speculation as to the network's true motivations for dumping the show. Barris has commented that he heard that NBC's official reason was because of both "lower than expected ratings" and a desire by the network to "re-tailor the morning shows to fit the standard morning demographics" (the move coincided with the arrival of new NBC president Fred Silverman, who was well known for such programming overhauls and was reported to have disliked The Gong Show). America Alive!, a magazine-style variety program hosted by Art Linkletter's son Jack, replaced Gong.
Following the cancellation, many critics and industry analysts – including Gene Shalit and Rona Barrett – reported having heard comments from within the NBC programming department from "sources preferring anonymity" that the true reason behind the cancellation was Barris's refusal to tone down the increasingly racy nature of the show. According to the sources, after the "Popsicle Twins" incident and an episode in which Jaye P. Morgan exposed her breasts on air during a Gene Gene the Dancing Machine segment, Barris had been given an ultimatum by the network's Standards and Practices department to deliver cleaner shows for his audience, which included many younger viewers, or NBC would cancel the program.
Finale
NBC allowed Barris to continue the show for the rest of his contract, and Barris made no perceptible change in preparation for the finale. In the finale, staff member Larry Gotterer appeared as "Fenwick Gotterer" to host the show after Barris started the show doing a "Chuckie's Fables" sketch. The rest of the final episode tried to explain the life of the show and its cancellation. Barris managed to have the last word on the show's demise, appearing as a contestant. Playing in a country music band called "The Hollywood Cowboys" with the house band's rhythm section, Barris sang a slightly modified version of Johnny Paycheck's "Take This Job and Shove It", giving NBC the finger during the song to accentuate his point. NBC censored the gesture, with the word "OOPS!" superimposed over a still shot of the set. Barris was gonged by Jamie Farr, who quipped, "Because that little fella's been saying that I've been long of nose, I'm also long of gong, fella."
The group "Lobster Repair" (who performed Harry Belafonte's song "Day-O") won the final $516.32 and trophy of the NBC era.
Syndicated (1976–80)The Gong Show continued in syndication for two years after its daytime counterpart's cancellation, often airing on weekends and at night. The entire syndicated run from September 1976 to September 1980 was distributed by Firestone Program Services. While the series eventually met its demise in syndication as it had on NBC, according to Barris, the problem did not lie with any outrageous acts, but instead the controversy and public outcry over another series he had produced.
In September 1979, Barris launched the game show Three's a Crowd, which was a spin-off of The Newlywed Game. Instead of recently married couples trying to match answers, the wives and secretaries of married men would compete to show who knew the men better. Religious activists and feminist groups protested against Three's a Crowd and its ratings eventually forced the show's cancellation during the middle of the season.
In Barris' autobiography The Game Show King, he wrote that "the public backlash from Three's a Crowd not only caused the program to be canceled, but it took three other TV shows of mine with it. I went to my house in Malibu and stayed there for a year." Gong was one of those shows to be canceled, and Barris never hosted another series. The trauma from the Three's a Crowd's backlash was so severe, in the last several weeks of the Gong Show, Barris reportedly had "a small nervous breakdown" on-air, because he was "bored to death" with broadcasting. His next two series, revivals of the 1960s game show Camouflage (the replacement for Three's a Crowd) and his 1973–77 series Treasure Hunt (toward which Barris had little or no input, according to host Geoff Edwards), both failed to find audiences and Barris went further into his self-imposed exile from television. Barris would not have another hit series until the 1985 syndication revival of The Newlywed Game.
Reruns of the NBC airings began in syndication in Fall 1979. The NBC and syndicated episodes were rerun on the USA Network and Game Show Network, although by the time GSN picked up the series, many episodes could not be aired because of musical performance clearance issues. No episodes from the first syndicated season were rerun.
Later incarnations
A syndicated weekday revival of The Gong Show, hosted by San Francisco disc jockey Don Bleu, ran during the 1988–89 season from September 12, 1988 to September 15, 1989, Each winner was paid $701.Extreme Gong, a later incarnation of The Gong Show on the Game Show Network had viewers vote on its acts by telephone. It was hosted by George Gray and ran from October 5, 1998 to August 26, 1999, with reruns continuing to air up until Fall of 2000. Winners received $317.69. This version was well known for two known incidents: one episode featuring "Cody the Talking Dog" for which he tried to say things like "I love you" and "ice cream" but did not succeed in talking, and another featuring a Village People parody as The Village Little People where they sang a cover of "YMCA". Orange County comedy punk band the Radioactive Chicken Heads (then called Joe & the Chicken Heads) made their national television debut on Extreme Gong, though they were gonged midway through their performance. Near the end of the show's run, an hour-long "Tournament of Talent" special was aired in August 1999, with twelve previous winning acts (chosen by viewers via a phone-in poll) competing for a payoff of $10,000.
Comedy Central debuted a new incarnation called The Gong Show with Dave Attell, which lasted for eight weeks in the summer of 2008. The show's format was similar to the original, but its scoring was based on a scale of 0 to 500, and winning acts received $600. The $600 was shown as paid in cash on the spot, rather than being paid by check as in earlier versions, but in reality (because of contestant eligibility regulations by Sony) was paid as a check from Sony Pictures. In place of a typical trophy, winners were awarded a belt in the style of boxing championship belts.
A live stage version of The Gong Show took place at B.B. King's Blues Club, in the Times Square district of New York City on August 12, 2010. It was produced by The Radio Chick, and is the Sony authorized stage production. This production went into development in 2011–12 and now runs regularly in New York City, with engagements in other U.S. cities.
2017 revival
On October 3, 2016, ABC and Sony Pictures announced a 10-episode summer 2017 revival of The Gong Show executive produced by Will Arnett.
The broadcast of the 2017 revival premiered on ABC on June 22, 2017, hosted by previously-unknown "British comedian" Tommy Maitland. Maitland is, in fact, a character portrayed by Mike Myers, although neither ABC nor Myers confirmed this and ABC officially credited Maitland as host and executive producer. Maitland's catchphrase is "Who's a cheeky monkey?" He also periodically uses Barris' "back with more stuff" catchphrase to lead into commercials.
Celebrity guest judges for the 2017 revival included Arnett, Zach Galifianakis, Alison Brie, Andy Samberg, Elizabeth Banks, Tracee Ellis Ross, Joel McHale, Megan Fox, Courteney Cox, Dana Carvey, Will Forte, Jack Black, Jennifer Aniston, Ken Jeong, Fred Armisen, Maya Rudolph and Anthony Anderson. Among the more notable acts featured on the revival are the Radioactive Chicken Heads, making their second appearance on a Gong Show incarnation since Extreme Gong in 1998. The first season of this version also features a regular segment featuring a staff performer leading the audience in a sing-along of the novelty song "Shaving Cream", reminiscent of the recurring gag acts on the earlier version.
The winner of each show received a gong trophy and an oversized check in the amount of US$2,000.17, later increased by a penny the following year. The final episode of the first season featured a memorial to Barris, who died prior to the series' premiere in 2017.
On January 8, 2018, ABC announced that the revival would be picked up for a second season, officially confirmed Myers as the portrayer of Maitland and crediting Myers as an executive producer by name. The second (and final) season premiered on June 21, 2018. Celebrity guest judges for season 2 included Jimmy Kimmel, Jason Sudeikis, Brad Paisley, Dana Carvey, Alyson Hannigan, Ken Jeong, Kristen Schaal, and Rob Riggle. However, the series was not renewed for a third season and was quietly cancelled.
Film
In 1980, The Gong Show Movie was released by Universal Pictures to scathing reviews and was quickly withdrawn from theatrical release. Advertising proclaimed it as "The Gong Show That Got Gonged by the Censor". It is seen periodically on cable TV, but was not released on home video until March 29, 2016, when the film was released on Blu-ray by Shout! Factory.Confessions of a Dangerous Mind, a film directed by George Clooney and written by Charlie Kaufman, was based on the semi-fictional autobiography of the same name by Chuck Barris. Part of the film chronicles the making of The Gong Show, and features several clips from the original series.
Following the success of the print and screen versions of Confessions, GSN produced a documentary called The Chuck Barris Story: My Life on the Edge, which included rare footage from the Gary Owens pilot.
International versions
Spinoffs
At the height of the Gong Show's popularity, NBC gave Barris a prime-time variety hour, The Chuck Barris Rah Rah Show. This was played somewhat more seriously than the Gong Show, with Jaye P. Morgan singing straight pop songs as in her nightclub and recording days, and bygone headliners like Slim Gaillard reprising their old hits for a studio audience. Other spinoffs include The $1.98 Beauty Show hosted by Rip Taylor and The Gong Show Movie.
See also
Minutes to Fame'', a similar Hong Kong talent show
References
External links
ABC version
The Gong Show (1976) on IMDb
The Gong Show (1988) on IMDb
Extreme Gong on IMDb
The Gong Show with Dave Attell on IMDb
The Gong Show (2017) on IMDb
American game shows
1970s American comedy game shows
1976 American television series debuts
1980 American television series endings
1980s American comedy game shows
1988 American television series debuts
1989 American television series endings
1990s American comedy game shows
2000s American comedy game shows
2010s American comedy game shows
2017 American television series debuts
2018 American television series endings
American Broadcasting Company original programming
American television series revived after cancellation
English-language television shows
First-run syndicated television programs in the United States
NBC original programming
Talent shows
Television series by Barris Industries
Television series by Sony Pictures Television |
null | null | Silent majority | eng_Latn | The silent majority is an unspecified large group of people in a country or group who do not express their opinions publicly. The term was popularized by U.S. President Richard Nixon in a televised address on November 3, 1969, in which he said, "And so tonight—to you, the great silent majority of my fellow Americans—I ask for your support." In this usage it referred to those Americans who did not join in the large demonstrations against the Vietnam War at the time, who did not join in the counterculture, and who did not participate in public discourse. Nixon, along with many others, saw this group of Middle Americans as being overshadowed in the media by the more vocal minority.
Preceding Nixon by half a century, it was employed in 1919 by Calvin Coolidge's campaign for the 1920 presidential nomination. Before that, the phrase was used in the 19th century as a euphemism referring to all the people who have died, and others have used it before and after Nixon to refer to groups of voters in various nations of the world.
Early meanings
Euphemism for the dead
'The majority' or 'the silent majority' can be traced back to the Roman writer Petronius, who wrote abiit ad plures (he is gone to the majority) to describe deceased people, since the dead outnumber the living. (In 2011 there were approximately 14 dead for every living person.). The phrase was used for much of the 19th century to refer to the dead. Phrases such as "gone to a better world", "gone before", and "joined the silent majority" served as euphemisms for "died". In 1902, Supreme Court Justice John Marshall Harlan employed this sense of the phrase, saying in a speech that "great captains on both sides of our Civil War have long ago passed over to the silent majority, leaving the memory of their splendid courage."
Groups of voters
In May 1831, the expression "silent majority" was spoken by Churchill C. Cambreleng, representative of New York state, before 400 members of the Tammany Society. Cambreleng complained to his audience about a U.S federal bill that had been rejected without full examination by the United States House of Representatives. Cambreleng's "silent majority" referred to other representatives who voted as a bloc:
In 1883, an anonymous author calling himself "A German" wrote a memorial to Léon Gambetta, published in The Contemporary Review, a British quarterly. Describing French Conservatives of the 1870s, the writer opined that "their mistake was, not in appealing to the country, but in appealing to it in behalf of a Monarchy which had yet to be defined, instead of a Republic which existed; for in the latter case they would have had the whole of that silent majority with them."
In 1919, Madison Avenue advertising executive and Republican Party supporter Bruce Barton employed the term to bolster Calvin Coolidge's campaign for the 1920 Republican Presidential nomination. In Collier's magazine, Barton portrayed Coolidge as the everyman candidate: "It sometimes seems as if this great silent majority had no spokesman. But Coolidge belongs with that crowd: he lives like them, he works like them, and understands."
Referring to Charles I of England, historian Veronica Wedgwood wrote this sentence in her 1955 book The King's Peace, 1637–1641: "The King in his natural optimism still believed that a silent majority in Scotland were in his favour."
Nixon
Also in 1955, while Nixon was serving as vice-president to Dwight D. Eisenhower, John F. Kennedy and his research assistants wrote in his book Profiles in Courage, "Some of them may have been representing the actual sentiments of the silent majority of their constituents in opposition to the screams of a vocal minority..." In January 1956, Kennedy gave Nixon an autographed copy of the book. Nixon wrote back the next day to thank him: "My time for reading has been rather limited recently, but your book is first on my list and I am looking forward to reading it with great pleasure and interest." Nixon wrote Six Crises, some say his response to Kennedy's book, after visiting Kennedy at the White House in April 1961.
In 1967, labor leader George Meany asserted that those labor unionists (such as himself) who supported the Vietnam War were "the vast, silent majority in the nation." Meany's statement may have provided Nixon's speechwriters with the specific turn of phrase.
Barbara Ehrenreich and Jay Caspian Kang later argued that awareness by the media and politicians that there actually might be a silent majority opposed to the anti-war movement was heightened during the August 1968 Democratic National Convention in Chicago, especially in reaction to the widely broadcast violence by police against protesters and media there. The media reacted indignantly "against the police and the mayor" after journalists and protesters were attacked and beaten by the police, but were stunned to find that a poll showed 56% of those surveyed "sympathized with the police". "Overnight the press abandoned its protest", awaking "to the disturbing possibility that they had grown estranged from a sizable segment of the public."
In the months leading up to Nixon's 1969 speech, his vice-president Spiro T. Agnew said on May 9, "It is time for America's silent majority to stand up for its rights, and let us remember the American majority includes every minority. America's silent majority is bewildered by irrational protest..." Soon thereafter, journalist Theodore H. White analyzed the previous year's elections, writing "Never have America's leading cultural media, its university thinkers, its influence makers been more intrigued by experiment and change; but in no election have the mute masses more completely separated themselves from such leadership and thinking. Mr. Nixon's problem is to interpret what the silent people think, and govern the country against the grain of what its more important thinkers think."
On October 15, 1969, the first Moratorium to End the War in Vietnam demonstrations were held, attracting thousands of protesters. Feeling very much besieged, Nixon went on national television to deliver a rebuttal speech on November 3, 1969, where he outlined "my plan to end the war" in Vietnam. In his speech Nixon stated his policy of Vietnamization would lower American losses as the South Vietnamese Army would take on the burden of fighting the war; announced his willingness to compromise provided that North Vietnam recognized South Vietnam; and finally promised he would take "strong and effective measures" against North Vietnam if the war continued. Nixon also implicitly conceded to the anti-war movement that South Vietnam was really not very important as he maintained that the real issue was the global credibility of the United States as he stated his belief that all of America's allies would lose faith in American promises if the United States were to abandon South Vietnam. Nixon ended his speech by saying all of this would take time, and asked for the public to support his policy of winning "peace with honor" in Vietnam as he concluded: "And so tonight, to you, the great silent majority of my fellow Americans—I ask for your support. Let us be united for peace. Let us be united against defeat. Because let us understand: North Vietnam cannot defeat or humiliate the United States. Only Americans can do that". The public reaction to the "silent majority speech" was very favorable at the time and the White House phone lines were overwhelmed with thousands of phone calls in the hours afterward as too many people called to congratulate the president for his speech.
Thirty-five years later, Nixon speechwriter Pat Buchanan recalled using the phrase in a memo to the president. He explained how Nixon singled out the phrase and went on to make use of it in his speech: "We [had] used 'forgotten Americans' and 'quiet Americans' and other phrases. And in one memo I mentioned twice the phrase 'silent majority,' and it's double-underlined by Richard Nixon, and it would pop up in 1969 in that great speech that basically made his presidency." Buchanan noted that while he had written the memo that contained the phrase, "Nixon wrote that speech entirely by himself."
Nixon's constituency
Nixon's silent majority referred mainly to the older generation (those World War II veterans in all parts of the U.S.) but it also described many young people in the Midwest, West and in the South, many of whom eventually served in Vietnam. The Silent Majority was mostly populated by blue collar white people who did not take an active part in politics: suburban, exurban and rural middle class voters. They did, in some cases, support the conservative policies of many politicians.
According to columnist Kenneth Crawford, "Nixon’s forgotten men should not be confused with Roosevelt's," adding that "Nixon's are comfortable, housed, clad and fed, who constitute the middle stratum of society. But they aspire to more and feel menaced by those who have less."
In his famous speech, Nixon contrasted his international strategy of political realism with the "idealism" of a "vocal minority." He stated that following the radical minority's demands to withdraw all troops immediately from Vietnam would bring defeat and be disastrous for world peace. Appealing to the silent majority, Nixon asked for united support "to end the war in a way that we could win the peace." The speech was one of the first to codify the Nixon Doctrine, according to which, "the defense of freedom is everybody's business—not just America's business." After giving the speech, Nixon's approval ratings which had been hovering around 50% shot up to 81% in the nation and 86% in the South.
In January 1970, Time put on their cover an abstract image of a man and a woman representing "Middle America" as a replacement for their annual "Man of the Year" award. Publisher Roy E. Larsen wrote that "the events of 1969 transcended specific individuals. In a time of dissent and 'confrontation', the most striking new factor was the emergence of the Silent Majority as a powerfully assertive force in U.S. society." Larsen described how the silent majority had elected Nixon, had put a man on the moon, and how this demographic felt threatened by "attacks on traditional values".
The silent majority theme has been a contentious issue amongst journalists since Nixon used the phrase. Some thought Nixon used it as part of the Southern strategy; others claim it was Nixon's way of dismissing the obvious protests going on around the country, and Nixon's attempt to get other Americans not to listen to the protests. Whatever the rationale, Nixon won a landslide victory in
1972, taking 49 of 50 states, vindicating his "silent majority". The opposition vote was split successfully, with 80% of George Wallace supporters voting for Nixon rather than George McGovern, unlike Wallace himself.
Nixon's use of the phrase was part of his strategy to divide Americans and to polarize them into two groups. He used "divide and conquer" tactics to win his political battles, and in 1971 he directed Agnew to speak about "positive polarization" of the electorate. The "silent majority" shared Nixon's anxieties and fears that normalcy was being eroded by changes in society. The other group was composed of intellectuals, cosmopolitans, professionals and liberals, those willing to "live and let live." Both groups saw themselves as the higher patriots. According to Republican pollster Frank Luntz, "silent majority" is but one of many labels which have been applied to the same group of voters. According to him, past labels used by the media include "silent majority" in the 1960s, "forgotten middle class" in the 1970s, "angry white males" in the 1980s, "soccer moms" in the 1990s, and "NASCAR dads" in the 2000s.
Later use
"Silent majority" was the name of a movement (officially called Anticommunist City Committee) active in Milan, Italy, from 1971 to 1974 and headed by the former monarchist partisan Adamo Degli Occhi, that expressed the hostility of the middle class to the 1968 movement. At the beginning it was of conservative tendency; later it moved more and more to the right and in 1974 Degli Occhi was arrested because his relationships with the terroristic movement Movimento di Azione Rivoluzionaria (MAR).
In 1975, in Portugal, then president António de Spínola used the term in confronting the more radical forces of post-revolutionary Portugal.
The phrase "silent majority" has also been used in the political campaigns of Ronald Reagan during the 1970s and 1980s, the Republican Revolution in the 1994 elections, and the victories of Rudy Giuliani and Michael Bloomberg. The phrase was also used by Quebec Premier Jean Charest during the 2012 Student Strike to refer to what he perceived as the majority of the Quebec voters supporting the tuition hikes.
The term was used by British Prime Minister David Cameron during the 2014 Scottish independence referendum; Cameron expressed his belief that most Scots opposed independence, while implicitly conceding they may not be as vocal as the people who support it.
During Donald Trump's presidential campaign, he said at a campaign rally on July 11, 2015, in Phoenix, Arizona, that "the silent majority is back, and we’re going to take our country back". He also referred to the silent majority in subsequent speeches and advertisement, as did the press when describing those who voted for his election as President in 2016. In the midst of the George Floyd protests, he once again invoked the silent majority. CNN analyst Harry Enten described that Trump's support fits better with the term "loud minority", based on the fact that he never hit 50% in any live interview opinion poll throughout his presidency. Jay Caspian Kang argues that some politicians and analysts (Jim Clyburn, Chuck Rocha) feel the unexpected increase in support for Donald Trump among blacks and Latinos in the 2020 election reflects a new silent majority (including some non-whites) reacting against calls for defunding the police and the arrogance of “woke white consultants”.
In 2019, the Prime Minister of Australia, Scott Morrison, acknowledged the quiet Australians in his federal election victory speech.
In the face of rising opposition, the Hong Kong government often claims there is a silent majority that is too afraid to voice their support, and a group called "Silent Majority for Hong Kong" was set up in 2013 to counteract the Occupy Central with Love and Peace movement. In 2019, when the democratic movement became increasingly violent, the Carrie Lam administration and Beijing authorities appealed to the "silent majority" to dissociate themselves from the radical activists and to vote for the pro-government camp in the District Council elections, which were seen as a de facto referendum on the protests. However, with a record turnout of over 70%, the pro-democracy camp won 80% of overall seats and controlled 17 out of the 18 District Councils. A commentator of The New Statesman deduced that Hong Kong's true silent majority stood on the side of the democratic cause. Foreign Policy stated that Beijing had been confident of a huge pro-government victory as a result of a delusion created by its own propaganda.
See also
Bradley effect
Majority rule
Democracy
Majoritarianism
Visible minority
Spiral of silence
Shy Tory Factor
Mainstream media
Pact of forgetting
Social desirability bias
1% rule (Internet culture)
References
Further reading
Browne, Junius Henri (1874). "The Silent Majority". Harper's Magazine, June to November
Campbell, Karlyn Kohrs. The Great Silent Majority: Nixon's 1969 Speech on Vietnamization (Texas A&M University Press; 2014) focus on the speech of November 3, 1969
American political catchphrases
Presidency of Richard Nixon
Majority
1830s neologisms |
null | null | NAMES Project AIDS Memorial Quilt | eng_Latn | The NAMES Project AIDS Memorial Quilt, often abbreviated to AIDS Memorial Quilt or AIDS Quilt, is an enormous memorial to celebrate the lives of people who have died of AIDS-related causes. Weighing an estimated 54 tons, it is the largest piece of community folk art in the world as of 2020.
History and structure
The idea for the NAMES Project Memorial Quilt was conceived on November 27, 1985, by AIDS activist Cleve Jones during the annual candlelight march, in remembrance of the 1978 assassinations of San Francisco Supervisor Harvey Milk and Mayor George Moscone. For the march, Jones had people write the names of loved ones that were lost to AIDS-related causes on signs, and then they taped the signs to the old San Francisco Federal Building. All the signs taped to the building looked like an enormous patchwork quilt to Jones, and he was inspired. The NAMES Project officially started in 1987 in San Francisco by Jones, Mike Smith, and volunteers Joseph Durant, Jack Caster, Gert McMullin, Ron Cordova, Larkin Mayo, Steve Kirchner, and Gary Yuschalk. At that time many people who died of AIDS-related causes did not receive funerals, due to both the social stigma of AIDS felt by surviving family members and the outright refusal by many funeral homes and cemeteries to handle the deceased's remains. Lacking a memorial service or grave site, The Quilt was often the only opportunity survivors had to remember and celebrate their loved ones' lives. The first showing of The Quilt took place on October 11, 1987 on the National Mall in Washington, DC, as part of the National March on Washington for Lesbian and Gay rights. The Quilt was last displayed in full on the Mall in Washington, D.C., in 1996, a display that included a visit by President Bill Clinton and first lady Hillary Rodham Clinton, but it returned in July 2012 to coincide with the start of the XIX International AIDS Conference.
The Quilt is a memorial to and celebration of the lives of people lost to the AIDS pandemic which marks it as a prominent forerunner of the twentieth century shift in memorial design that moved towards celebrating victims or survivors. Each panel is by , approximately the size of the average grave; this connects the ideas of AIDS and death more closely. The Quilt is still maintained and displayed by The NAMES Project Foundation.
In observance of National HIV-Testing Day in June 2004 the 1,000 newest blocks were displayed by the Foundation on The Ellipse in Washington, D.C. The largest display of The Quilt since it was last displayed in its entirety in October 1996, the 1,000 blocks displayed consisted of every panel submitted at or after the 1996 display.
In 1997, the NAMES Project headquarters moved from San Francisco to Washington, D.C., and in 2001 the quilt panels were moved from San Francisco to Atlanta, Georgia. The NAMES Project Foundation was headquartered in Atlanta.
In 2019, the organization announced that the Quilt would be relocating to San Francisco under the care of the National AIDS Memorial. Its archives would relocate to the American Folklife Center at the Library of Congress. The AIDS Memorial Quilt is now warehoused in San Francisco when not being displayed, and continues to grow, currently consisting of more than 48,000 individual memorial panels (over 94,000 people) and weighing an estimated 54 tons.
Goal and achievement
The goal of the Quilt is to bring awareness to how massive the AIDS pandemic really is, and to bring support and healing to those affected by it. Another goal is to raise funds for community-based AIDS service organizations, to increase their funding for AIDS prevention and education. As of 1996, more than $1.7 million had already been raised, and the effort continues to this day.
Quilt construction and care
panels made typically of fabric are created in recognition of a person who died from AIDS-related complications. The panels are made by individuals alone or in a workshop, such as Call My Name (which focuses on African American representation on the quilt) or in quilting bees, such as the one held during the 2012 Smithsonian Folklife Festival on the National Mall. Construction choices are left to the quilter and techniques such as traditional fabric quilting, embroidery, applique, paint and stencil, beading, and iron-ons are common.
Items and materials included in the panels:
Fabrics such as lace, suede, leather, mink, taffeta, also Bubble Wrap and other kinds of plastic and even metal.
Decorative items like pearls, quartz crystals, rhinestones, sequins, feathers, buttons.
Clothing, such as jeans, T-shirts, gloves, boots, hats, uniforms, jackets, flip-flops.
Items of a personal nature, such as human hair, cremation ashes, wedding rings, merit badges and other awards, car keys.
Unusual items, such as stuffed animals, records, jockstraps, condoms, and bowling balls.
Panels are submitted to the National AIDS Memorial, along with a panel-maker identification form and a documentation letter. Occasionally, other supplemental material is donated along with the panel such as photographs of the subject. The information about the panel is recorded in a database.
Panels are backed in canvas and sewn together in blocks of eight. Grommets for hanging are attached and the blocks are numbered and photographed. The numbers help with identification and location in storage, on the quilt website, and when the quilt is displayed.
Quilt maintenance and Gert McMullin
The quilt is maintained, repaired, and managed by Hand Maidens of the Quilt. The most dedicated Hand Maiden is Gert McMullin. McMullin, chief quilt production coordinator for the NAMES Project Foundation, volunteered her anger-driven energy and sewing skills to Cleve Jones in the early days of the quilt in San Francisco. After witnessing and experiencing the deaths of so many friends from HIV, McMullin dedicated herself, working nights after her job at a Macy's cosmetic counter, to combat the invisibility her community felt. Materials and sewing machines were donated and McMullin and a group of volunteers worked in a storefront on Market Street. They created hundreds and later thousands of panels.
McMullin's first two panels were for her friends, Roger Lyon and David Calgaro. Her panel for Lyon was eventually accessioned into the collection of the Smithsonian Institution National Museum of American History (accession number 1998.0254.01) and was featured in the book The Smithsonian's History of America in 101 Objects.
When San Francisco became prohibitively expensive, the project, and McMullin, moved to Atlanta, Georgia. McMullin returned to San Francisco in 2020 as the Quilt returned to the city where it originated under the care and stewardship of the National Aids Memorial. Gert continues to care for each Quilt panel and works with a team of Quiltmakers and caretakers to maintain and preserve each panel and organize displays around the country.
During the COVID-19 pandemic, Gert and her colleagues struck by the similarities of the COVID and AIDS pandemics, wanted to help those in need. They began making protective masks made from quilting fabric for first responders as a way to show their support and deal with their own grief and memories of loved ones being lost during the AIDS pandemic. They have made thousands of protective masks and delivered them to organizations throughout the San Francisco Bay Area.
Examples of panels
Those who submit panels do not have to know the person, but they do have to feel some sort of connection with the individual that they want people to recognize. For example, to memorialize Queen lead-singer Freddie Mercury, there were many panels made, two of which were a solid white background with a blue and black guitar, and "Freddy Mercury" written down the sides in black, with the AIDS ribbon above his name, and a purple silk with "Freddie Mercury", "Queen", and "1946–1991" in silver applique, along with two pictures of Mercury with Queen.
Many panels were also made for the actor Rock Hudson, one of which consisted of a navy blue background with silver "Rock Hudson" and stars, above a rainbow with the word "Hollywood".
Other panels are made by loved ones and then attached to make one large block. Some are flamboyant and loud, whereas some are more muted and simple.
Recognition and influence
The NAMES Project was nominated for a Nobel Peace Prize in 1989.
The Quilt is the subject of the 1989 Peabody Award- and Academy Award-winning documentary film, Common Threads: Stories from the Quilt, produced by Rob Epstein and Bill Couturié, and narrated by Dustin Hoffman.
Songwriter Tom Brown wrote the song "Jonathan Wesley Oliver, Jr." about the Quilt in 1988.
In 1990, John Corigliano's Symphony No. 1, inspired by The AIDS Memorial Quilt, premiered in New York.
Elegies for Angels, Punks and Raging Queens, a song cycle developed in the late 1980s with music by Janet Hood and lyrics and additional text by Bill Russell, features songs and monologues inspired by The Quilt.
In 1992 The AIDS Quilt Songbook premiered, a collection of new musical works about the devastation of AIDS compiled by lyric baritone William Parker, who solicited them from composers with whom he had previously worked.
Washington D.C.'s Different Drummers (DCDD) and the Lesbian and Gay Chorus of Washington (LGCW) commissioned Quilt Panels from composer Robert Maggio, and the piece premiered in 2003.
The NAMES Project was the basis for the musical Quilt, A Musical Celebration.
The AIDS Memorial Quilt was mentioned and shown during the years that General Hospital held their Nurses Ball (1994–2001) and raised money for AIDS research. The show's fictional character, Michael "Stone" Cates, was celebrated with a quilt block in 1996.
In 2002 the NAMES Project Chicago Chapter was inducted into the Chicago Gay and Lesbian Hall of Fame.
Never to Be Forgotten is a Philo T. Farnsworth Award-winning 54-minute video created by Karen Peper which documents the Quilt's June 1988 visit to Detroit, Michigan. This display was part of a 20-city tour initiated immediately after the 1987 Washington, DC inaugural showing. The video begins with footage of the opening ceremony from the Washington DC display and then moves to coverage of the Detroit event. Included are the opening and closing ceremonies at Cobo Hall along with a look at the set up and take down of the display. Volunteers share their feelings about participating in the event and the viewer is given a close-up look at the individual panels. Peper also shot extensive footage of the Quilt's visit to Columbus, OH; Chicago, IL; and the 1987, 1993, and 1996 Washington, D.C. showings. (All video footage is archived at ONE Archives at the USC Libraries in Los Angeles, CA.)
The Quilt was displayed on the ABC Soap Opera "One Life to Live" in 1992.
On Inside Amy Schumer, Schumer joked that comedian Mindy Kaling wore the quilt to an event during a skit mocking sizes in clothing stores.
On Modern Family season 7 episode 1, Luke joked Haley's quilt was the saddest since the AIDS Quilt.
Cartoonist Gerard Donelan, specializing in single-panel comics depicting gay men and women in everyday life, contributed cartoons, pamphlets, and posters to the NAMES Project. These pieces were meant to spread awareness about safe-sex practices for gay people and to garner support for the NAMES Project.
Projects inspired by NAMES
The AIDS Memorial Quilt was the first of its kind as a continually growing monument created piecemeal by thousands of individuals, and today it constitutes the largest piece of community folk art in the world. The Quilt was followed by, and inspired a number of memorials and awareness projects, both AIDS-related and otherwise. Examples of these include:
The K.I.A. Memorial Quilt, created to remember those U.S. Armed Forces members killed in the Iraq War.
Following the September 11, 2001 terrorist attacks on America a number of quilt projects were created memorializing the victims.
September 11 Quilts Memorial Exhibition
United In Memory
The World Trade Center Memorial Quilt
America's 9-11 Memorial Quilts
Many other medical conditions also now have quilts, for example:
Huntington's disease
Congenital heart disease
Breast cancer
There are also quilts for sub-sects of the AIDS Pandemic, including:
Children
North Californians
Australians
New Zealand
"Virtual" AIDS Memorial Quilts have also been created:
Project Stitch "Digital Quilt"
Second Life
AIDS Action Committee of Massachusetts
Southern AIDS Living Quilt
Columbia University AIDS Memorial
Display location
In November 2019 the NAMES Project Foundation and House Speaker Nancy Pelosi announced that the quilt would relocate to San Francisco under the permanent care and stewardship of the National AIDS Memorial starting in 2020. The Project's archives were gifted to the joint care with the American Folklife Center at the U.S. Library of Congress, allowing for greater public access. This action returns the quilt to San Francisco, where the project began.
See also
Art of the AIDS Crisis
New York City AIDS Memorial
References
Further reading
New York Memorial Quilt records, 1988 (3 inches) are housed at the New York State Archives.
Robert Garcia Papers, 1988–1993 (9 cubic feet) are housed at the Cornell University Library Division of Rare and Manuscript Collection.
Jim Graham Papers, 1961-2015 (80.5 linear feet) are housed at the George Washington University Special Collections Research Center.
Vito Russo Papers, 1969–1990 (21 linear feet) are housed at the New York Public Library.
Washington Blade Lou Chibbaro Senior Report files, 1980–2001 (39 linear feet) are housed at the George Washington University Special Collections Research Center..
2362 Market Street Stories by Patricia Wakida, 2012 Smithsonian Folklife Festival, Creativity in Crisis, blog profile of Gert McMullin
Dupré, Judith (2007). Monuments: America's History in Art and Memory. New York: Random House.
External links
The NAMES Project Foundation AIDS Memorial Quilt website — images of the quilt.
National AIDS Memorial website
Documentary and overview description of the 1996 quilt display
"National Coalition to Save Our Mall: The AIDS Memorial Quilt" (archived)
HIV/AIDS activism
Monuments and memorials in the United States
Organizations based in Atlanta
Tourist attractions in Atlanta
Quilts
1987 works
LGBT culture in Atlanta
History of San Francisco
1987 establishments in California
1987 in San Francisco
HIV/AIDS memorials
HIV/AIDS |
null | null | Flag of South Africa | eng_Latn | The flag of South Africa was designed in March 1994 and adopted on 27 April 1994, at the beginning of South Africa's 1994 general election, to replace the flag that had been used since 1928.
The flag has horizontal bands of red (on the top) and blue (on the bottom), of equal width, separated by a central green band which splits into a horizontal "Y" shape, the arms of which end at the corners of the hoist side (and follow the flag's diagonals). The "Y" embraces a black isosceles triangle from which the arms are separated by narrow yellow or gold bands; the red and blue bands are separated from the green band and its arms by narrow white stripes. The stripes at the fly end are in the 5:1:3:1:5 ratio. Three of the flag's colours were taken from the flag of the South African Republic, itself derived from the flag of the Netherlands, as well as the Union Jack, while the remaining three colours were taken from the flag of the African National Congress.
Colours
At the time of its adoption, the South African flag was the only national flag in the world to comprise six colours in its primary design and without a seal and brocade. The design and colours are a synopsis of principal elements of the country's flag history. The colours themselves have no essential meaning.
The central design of the flag, beginning at the flagpost in a "Y" form and flowing into a single horizontal band to the outer edge of the fly.
Construction sheet
Symbolism
According to official South African government information, the South African flag is "a synopsis of principal elements of the country's flag history." Although different people may attribute personal symbolism to the individual colours or colour combinations, "no universal symbolism should be attached to any of the colours." The only symbolism in the flag is the V or Y shape, which can be interpreted as "the convergence of diverse elements within South African society, taking the road ahead in unity".
From time to time explanations of the meanings or symbolism of the flag's colours are published in various media, including official government publications and speeches by government officials.
Three of the colours — black, green and gold — are found in the flag of the African National Congress. The other three — red, white and blue — are used in the modern flag of the Netherlands and the flag of the United Kingdom; the colours white and blue were also found in the old flag of South Africa.
Former South African President F.W. de Klerk, who proclaimed the new flag on 20 April 1994, stated in his autobiography, The Last Trek: a New Beginning, that chilli red was chosen instead of plain red (which Anglo-Africans would have preferred) or orange (as Afrikaners would have preferred).
History
1902–1910
The Anglo-Boer War between 1899 and 1902 ended with the Treaty of Vereeniging on 31 May 1902 and resulted in what is now South Africa falling under the British Union Flag. The former Boer Republics of the Orange Free State and the Zuid-Afrikaanse Republiek (Transvaal) became British colonies along with the existing Cape and Natal colonies. Each was also entitled to a colonial flag following in the British tradition.
1910–1928
On 31 May 1910 these four colonies came together to form the Union of South Africa and the individual colonial flags were no longer used and new South African flags came into being. Once again, as a British dominion the British Union Flag was to continue as the national flag and the standard British ensign pattern was used as a basis for distinctive South African flags.
As was the case throughout the British Empire, the Red and Blue Ensigns were the official flags for merchant and government vessels at sea, and the British Admiralty authorised them to be defaced in the fly with the shield from the South African coat of arms. These ensigns were not intended to be used as the Union's national flag, although they were used by some people as such. Although these ensigns were primarily intended for maritime use, they were also flown on land.
These flags never enjoyed much popular support due to the animosities lingering after the Anglo-Boer War - the Afrikaner descendants of the Dutch settlers from the former Boer Republics saw it as a British flag, and the English-speakers saw it was not the Union Flag alone.
1928–1994 flag
Due to the lack of popularity of these flags, there were intermittent discussions about the desirability of a more distinctive national flag for South Africa after 1910, it was only after a coalition government took office in 1925 that a bill was introduced in Parliament to introduce a national flag for the Union. This provoked an often violent controversy that lasted for three years based on whether the British Union Flag should be included in the new flag design or not. The Natal Province even threatened to secede from the Union should it be decided to remove it.
Finally, a compromise was reached that resulted in the adoption of a separate flag for the Union in late 1927 and the design was first hoisted on 31 May 1928. The design was based on the so-called Van Riebeeck flag or "Prince's Flag" (Prinsenvlag in Afrikaans) that was originally the Dutch flag; it consisted of orange, white, and blue horizontal stripes. A version of this flag had been used as the flag of the Dutch East India Company (known as the VOC) at the Cape (with the VOC logo in the centre) from 1652 until 1795. The South African addition to the design was the inclusion of three smaller flags centred in the white stripe. The miniature flags were the British Union Flag (mirrored) towards the hoist, the flag of the Orange Free State hanging vertically in the middle and the Transvaal Vierkleur towards the fly. The position of each of the miniature flags is such that each has equal status. However, to ensure that the Dutch flag in the canton of the Orange Free State flag is placed nearest to the upper hoist of the main flag, the Free State flag must be reversed. The British Union Flag, which is nearest to the hoist and is thus in a more favoured position, is spread horizontally from the Free State flag towards the hoist and is thus also reversed. Although placed horizontally furthest from the hoist, to balance the British Union Flag, the Vierkleur is the only one of the miniature flags which is spread in the same direction as the main flag. This compensates for its otherwise less favourable position. In this arrangement, each of the miniature flags enjoy equal precedence. Note that the miniature flags of the Transvaal Republic and the Orange Free State both contain miniature flags of the Netherlands, while the miniature flag of the United Kingdom is a composition of the flags of England, Scotland and the Anglo-Irish people, making the old South African flag the only former national flag in the world containing five flags within three flags within a flag.
The choice of the Prinsenvlag (which was believed to be the first flag hoisted on South African soil by Jan van Riebeeck of the VOC) as the basis upon which to design the South African flag had more to do with compromise than Afrikaner political desires, since the Prinsenvlag was politically neutral, as it was no longer the national flag of any nation. A further element of this compromise was that the British Union Flag would continue to fly alongside the new South African national flag over official buildings. This dual flag arrangement continued until 1957 when the British Union Flag lost its official status per an Act of Parliament.
Following a referendum the country became a republic on 31 May 1961, but the design of the flag remained unchanged. However, there was intense pressure to change the flag, particularly from Afrikaners who still resented the fact that the British Union Flag was a part of the flag. In 1968, the then Prime Minister, John Vorster, proposed the adoption of a new flag from 1971, to commemorate the tenth anniversary of the declaration of a republic but this never materialised.
Current flag since 1994
The present South African national flag was first flown on 27 April 1994, the day of the 1994 election. However, the flag was first intended to be an interim flag only, and its design was decided upon only a week beforehand.
The choice of a new flag was part of the negotiation process set in motion when Nelson Mandela was released from prison in 1990. When a nationwide public competition was held in 1993, the National Symbols Commission received more than 7,000 designs. Six designs were shortlisted and presented to the public and the Negotiating Council, but none elicited enthusiastic support. A number of design studios were then contacted to submit further proposals, but these also did not find favour. Parliament went into recess at the end of 1993 without a suitable candidate for the new national flag.
In February 1994, Cyril Ramaphosa and Roelf Meyer, the chief negotiators of the African National Congress and the National Party government of the day respectively, were given the task of resolving the flag issue. A final design was adopted on 15 March 1994, derived from a design developed by the State Herald Fred Brownell, who had also claimed to have previously designed the Namibian flag. This interim flag was hoisted officially for the first time on the 27 April 1994, the day when the nation's first fully inclusive elections commenced which resulted in Nelson Mandela being inaugurated as South Africa's first democratically elected president on 10 May 1994. The flag was well received by most South Africans, though a small minority objected to it; hundreds of Afrikaner Volksfront members in Bloemfontein burned the flag in protest a few weeks before the April 1994 elections.
The proclamation of the new national flag by South African President F. W. de Klerk was only published on 20 April 1994, a mere seven days before the flag was to be inaugurated, sparking a frantic last-minute flurry for flag manufacturers. As stated in South Africa's post-apartheid interim constitution, the flag was to be introduced on an interim probationary period of five years, after which there would be discussion about whether or not to change the national flag in the final draft of the constitution. The Constitutional Assembly was charged with the responsibility of drafting the country's new constitution and had called for submissions, inter alia, on the issues of its various national symbols. It received 118 submissions recommending the retention of the new flag and 35 suggesting changes to it. Thus on 28 September 1995 it decided that the flag should be retained unchanged and accordingly it was included as Section One of the Constitution of South Africa which came into force in February 1997.
Proper display of the flag
The South African government published guidelines for proper display of the flag at designated flag stations, in Government Notice 510 of 8 June 2001 (Gazette number 22356). These rules apply only to official flag stations and not to the general public.
The Southern African Vexillological Association (SAVA), a non-official association for the study of flags, published their own guide for proper display of the flag in 2002. This guide has no official authority but was drawn up with generally accepted vexillological etiquette and principles in mind.
Official description
An addendum to the Transitional Executive Council agenda (April 1994) described the flag in pseudo-heraldic terms as follows:
The National flag shall be rectangular in the proportion of two in the width to three to the length; per pall from the hoist, the upper band red (chilli) and lower band blue, with a black triangle at the hoist; over the partition lines a green pall one fifth the width of the flag, fimbriated white against the red and blue, and gold against the black triangle at the hoist, and the width of the pall and its fimbriations is one third the width of the flag.
Schedule One of the Constitution of South Africa (1996) replaced the heraldic definition and described the flag in plain English as follows:
The national flag is rectangular; it is one and a half times as long as it is wide.
It is black, gold, green, white, chilli red and blue.
It has a green Y-shaped band that is one fifth as wide as the flag. The centre lines of the band start in the top and bottom corners next to the flag post, converge in the centre of the flag, and continue horizontally to the middle of the free edge.
The green band is edged, above and below in white, and towards the flag post end, in gold. Each edging is one fifteenth as wide as the flag.
The triangle next to the flag post is black.
The upper horizontal band is chilli red and the lower horizontal band is blue. These bands are each one third as wide as the flag.
See also
List of South African flags
Coat of arms of South Africa
National anthem of South Africa
National symbols of South Africa
Southern African Vexillological Association
References
Further reading
External links
Fred Brownell and the creation of the South African flag
Southern African Vexillological Association (SAVA)
South Africa
Flags of South Africa
National symbols of South Africa
South African heraldry
South Africa
South Africa
1994 establishments in South Africa |
null | null | Collingwood Football Club | eng_Latn | The Collingwood Football Club, nicknamed the Magpies or colloquially the Pies, is a professional Australian rules football club based in Melbourne, that was formed in 1892 in the suburb of Collingwood. The club played in the Victorian Football Association (VFA) before joining seven other teams in 1896 to found the breakaway Victorian Football League, today known as the Australian Football League (AFL). Originally based at Victoria Park, Collingwood now plays its home games at the Melbourne Cricket Ground and has its training and administrative headquarters at Olympic Park Oval and the Holden Centre.
Collingwood has played in a record 44 VFL/AFL Grand Finals (including rematches), winning 15, drawing two and losing 27 (also a record). Regarded as one of Australia's most popular sports clubs, Collingwood has attracted the second-highest attendance figures and television ratings of any professional football team in the nation.
The club's song, "Good Old Collingwood Forever", dates back to 1906, making it the oldest song currently used in the AFL. Its home guernsey consists of black and white stripes, matching the colours of the Australian magpie.
Historically, the club's biggest rivals have been cross-town clubs Carlton, Melbourne and Richmond. Starting in 1995, Collingwood has also enjoyed a healthy Anzac Day rivalry with Essendon and since the turn of the millennium, the club has developed smaller rivalries with West Coast and Brisbane.
Collingwood fields a reserves team in the Victorian Football League (formerly the VFA) and women's teams in the AFL Women's and VFL Women's competitions. It also owns and operates a netball team in the National Netball League.
History
Formation and early years
The Collingwood Football Club was established on 12 February 1892.
Collingwood played its first game in the Victorian Football Association (VFA) against Carlton on 7 May 1892. The club won the VFA Premiership in 1896.
In 1897, Collingwood, along with fellow VFA clubs Fitzroy, Melbourne, St Kilda, Carlton, Essendon, South Melbourne and Geelong split from the VFA and formed the Victorian Football League (VFL). Collingwood won its first premiership in 1902, defeating Essendon by 33 points in the 1902 VFL Grand Final.
1920s and 1930s: Four consecutive premierships
Collingwood was the most successful Victorian club of the 1920s and 1930s, appearing in 13 out of a possible 20 Grand Finals during the period. Collingwood were premiers six times during this time, including four consecutive premierships between 1927 and 1930, a VFL record, and two consecutive premierships in 1935 and 1936. The club's coach during this period was Jock McHale, who served as coach from 1912 to 1949. Collingwood also had three Brownlow Medallists during the period, with Syd Coventry winning in 1927, Albert Collier in 1929 and Harry Collier in 1930. The club's ruthlessly successful period later earned the club the nickname "The Machine". American journalist and author Sam Walker included the Machine team in his book The Captain Class, which listed some the author's greatest teams in the history of world sport.
The Collingwood team of 1927–30 not only achieved four straight premierships, but did so with a winning percentage of around 86% across the four seasons, and an average winning margin of about five goals. In 1929 they also became the only team in history to go through a home-and-away season undefeated. Collingwood remains the only club in the history of the VFL/AFL to have been declared premiers on four successive occasions.
1950s: Melbourne rivalry
In the 1950s, rival club Melbourne enjoyed an era of unprecedented success, winning five premierships in six years (the last coming in 1960, and having been runner up in 1954). Collingwood lost two Grand Finals to Melbourne in this decade, but bounced back to win premierships in 1953 and 1958. Collingwood's 1958 premiership is much cherished by the club as it prevented Melbourne from equalling Collingwood's record four premierships in a row.
The 1958 premiership was however to be Collingwood's last for 32 years, as the club was to suffer a string of Grand Final defeats in coming decades.
Collingwood and Melbourne play their rival match every year within the Queens Birthday Weekend.
1959–89: "Colliwobbles"
A string of eight Grand Final losses, often by narrow margins, between 1960 and 1981 gave rise to a perception that the club was prone to "choking", a phenomenon wittily dubbed "Colliwobbles". Whether this perception is accurate remains a subject of debate; having only won one and drawn one of its last six Grand Finals. Lou Richards ceremoniously buried the Colliwobbles at Victoria Park after the club's 1990 premiership.
1990–99: Long-awaited premiership and struggles
The 1990 premiership team, coached by Leigh Matthews and captained by Tony Shaw, had a one-sided grand final win against Essendon, the Magpies recording a 48-point victory and ending a 32-year premiership drought which included eight grand final losses and one draw. The sight of club great Darren Millane, who died in a car-crash one year later, holding the ball aloft in triumph at the final siren is one of the indelible images of the match.
After the drought-breaking premiership, the club lapsed into a state of decline for the remainder of the decade, culminating with the club's second wooden spoon in 1999. The Magpies returned to finals, though were quickly eliminated, in the 1992 season against St Kilda and in the 1994 AFL season against West Coast. Matthews left as head coach at the end of the 1995 season and was replaced at the start of the following year by 1990 premiership captain Tony Shaw, who had only retired from football 18 months earlier. Mid-table finishes under Shaw were achieved for the next two seasons, before poor results in 1998 and 1999 saw Shaw announce his resignation.
2000–11: The Malthouse era
Media personality, sports journalist and administrator Eddie McGuire was elected President in October 1998. He oversaw the installation of new head coach Michael Malthouse in October 1999, whose appointment proved to be a masterstroke in reviving the club on-field. Under Malthouse, the acquisition and emergence of players such as Paul Licuria, Alan Didak, Anthony Rocca and Nathan Buckley resulted in Collingwood quickly moving up the ladder in the 2000 AFL season and in the 2001 AFL season, only narrowly missing the finals in the latter year. Collingwood met reigning premiers Brisbane in the 2002 Grand Final and were regarded as massive underdogs, eventually falling just 9 points short of an improbable premiership. Buckley, the captain, became just the third player to win the Norm Smith Medal as best afield in the Grand Final despite being a member of the losing side. Despite a very successful home-and-away next season, they were again defeated by the Lions in the 2003 Grand Final, this time in thoroughly convincingly fashion.
Following those Grand Final losses, Collingwood struggled for the next two years, finishing 13th in 2004 and second-last in 2005; the latter meant Collingwood was eligible for a priority pick which the club used to recruit Dale Thomas. Collingwood made a return to the finals in 2006, finishing fifth, but were defeated by the Western Bulldogs by 41 points in its elimination final. A loss to late in the season was to cost them the double chance. The 2007 season saw them finish sixth on the ladder at season's conclusion, and in the finals they knocked out the grand finalists of the past two years, Sydney, in the elimination final and then West Coast in overtime at Subiaco Oval in the semi-final. Having earned a preliminary final against , Collingwood lost to the eventual premiers, by five points in one of the most memorable preliminary finals in over a decade. Nathan Buckley would announce his retirement at season's end after playing just five games in 2007 due to injury.
Collingwood finished eighth in the 2008 AFL season and were assigned an away final against at AAMI Stadium. After at one point trailing in the match, Collingwood went on to end Adelaide's season and earn a semi-final meeting against . Having defeated the Saints in both their regular season meetings, Collingwood lost convincingly, ending their 2008 season. The 2009 season saw Collingwood finish inside the top-four for the first time since 2003, but in the qualifying final were beaten by minor premiers St Kilda convincingly. Having won a second chance, Collingwood struggled against Adelaide for the second year in a row before John Anthony kicked the match-winning goal with a minute left to send them into another preliminary final meeting with Geelong. But the season ended abruptly for the Magpies, with a 73-point loss to Geelong.
In 2010, Collingwood finished as minor premiers, and after wins in the qualifying and preliminary finals, reached the first Grand Final against . The match finished as a draw, forcing the first grand final replay in 33 years. Collingwood won the replay by 56 points. Key defensive player Nick Maxwell captained the club to victory and midfielder Scott Pendlebury (who had already won his first of eventually three Anzac medals earlier in the year) was awarded the Norm Smith Medal. The club won a second consecutive minor premiership in 2011, and qualified for the Grand Final after a three-point victory against Hawthorn in the preliminary final. However, Collingwood was then beaten by Geelong by 38 points in the decider, after trailing by seven points at three-quarter time. Following the Grand Final loss, which also marked the end of the club's 2011 AFL season, Malthouse left Collingwood after deciding not to stay on as "director of coaching". Star midfielder Dane Swan won the 2011 Brownlow Medal with a then-record 34 votes. Malthouse would leave having coached the club to eight finals series and four grand finals in 12 years.
2012–2021: Coach Nathan Buckley
Nathan Buckley, regarded as one of Collingwood's greatest players, was appointed assistant coach under Malthouse for the 2010 and 2011 seasons, before assuming the head coaching position at the start of the 2012 season. Malthouse, who had been contracted to take on a "head of coaching" role, elected to leave the club rather than put Buckley in what he regarded as an awkward position. Under Buckley, Collingwood continued to be successful in the short term, qualifying inside the top-four in the 2012 season, before falling 26 points short in a preliminary final to eventual premiers the Sydney Swans at ANZ Stadium. The club qualified for finals once more in 2013, though were surprisingly eliminated in the first week by underdogs Port Adelaide at home. The result prompted the Magpies coaching staff to begin making radical changes to the club's playing list, which saw premiership players Heath Shaw, Sharrod Wellingham, Heritier Lumumba among others leave for other clubs or retire. Over the next four years, younger talent was drafted but the club's win-loss recorded continued to deteriorate. Collingwood failed to make finals from 2014 through to the end of the 2017 season, progressively sliding down the ladder each year. Buckley came under intense media pressure to resign or be sacked from his position, though club administrators elected to grant him a two-year extension to his contract in October 2017 after a broad-ranging internal review.
The emergence of new-generation players such as Taylor Adams, Adam Treloar and Jordan De Goey, alongside key talls Brodie Grundy and Mason Cox mixed well with veterans Pendlebury and Steele Sidebottom. Collingwood jumped from 13th in 2017 to 3rd in 2018, sensationally knocking out reigning premiers in the preliminary final before falling five points short after leading for most of the match against West Coast in the 2018 Grand Final, the senior team's 27th defeat in a Grand Final. Buckley's growth as a coach was partially credited for the rapid improvement. In 2019, Collingwood had another strong season, finishing fourth on the ladder, but they were unable to return to the Grand Final after a shattering four-point defeat to in the first preliminary final. In 2020, Collingwood finished 8th at the end of the home-and-away season.
The club made significant on-field and administrative changes in the late 2010s. It was a foundation member of the inaugural AFL Women's competition in 2017 and in the same year established the Collingwood Magpies Netball team, a division of the club competing in the professional National Netball League. Collingwood unveiled a new permanent logo at the end of the 2017 season, which was the club's 125th anniversary year.
"Do Better" report
In 2020, the club commissioned an independent review into claims of racism at the club. In February 2021, the report was leaked to journalists and revealed that "while claims of racism have been made across the AFL, there is something distinct and egregious about Collingwood’s history" and that "what is clear is that racism at the club has resulted in profound and enduring harm to First Nations and African players. The racism affected them, their communities, and set dangerous norms for the public." Collingwood President Eddie McGuire suggested that the report signalled for the club which was working towards addressing racism and that it "was not a racist club". Many criticised McGuire's response, including AFL CEO Gillon McLachlan, Héritier Lumumba, former Indigenous Collingwood player Tony Armstrong and a Victorian Senator, among others. McGuire later apologised for the remarks. On 4 February, 150 Collingwood players from the men's and women's teams penned an open letter apologising "to anyone who, through their association with our club, has been marginalised, hurt or discriminated against due to their race." First-grade footballer Darcy Moore said that the players were "humiliated and shocked" by the report's findings. McGuire stood down as President of the Collingwood Football Club on 9 February 2021, although he had initially wanted to see the year through for a seamless transition until being compelled to step down.
Buckley stepped down after Round 13 of the 2021 AFL season, and assistant coach Robert Harvey took over as the caretaker coach until the end of the season. Harvey focused on developing youth and letting them play, with Collingwood winning 2 out of their 9 remaining games.
2022–: Coach Craig McRae
In September 2021, Craig McRae was appointed as head coach of the club for the 2022 season and onwards.
Club symbols and identity
Guernsey
Throughout the club's history, Collingwood has worn a guernsey of black and white vertical stripes. The all white jumper, with the three black vertical stripes is the iconic strip that the club is most associated with. The current incarnation of the guernsey is mostly black, with white stripes on the front and lower half of the back, and white numbers. The main clash guernsey is the reverse of this: mostly white, with black stripes and black numbers, worn in away matches against clubs with a predominantly dark guernsey such as Fremantle and Port Adelaide. A secondary clash guernsey was introduced in 2011 and is used only in matches against North Melbourne due to similarity between the two uniforms. The alternate uniform is black with only two white stripes on each side instead of three.
Traditionally, Collingwood has worn a white guernsey with black stripes. The club switched to the black guernsey with white stripes in 2001.
Nike is the current manufacturer of the Magpies' apparel.
Collingwoods cultural reach and impact is far reaching as evidence by memberships, crowds, broadcast ratings and more recently, the emergence of influential digital media, such as the Pie Hard podcast.
Song
"Good Old Collingwood Forever" is the team song of the Collingwood Football Club. The lyrics were written by player Tom Nelson during Collingwood's 1906 tour of Tasmania, making it the oldest of the team songs currently used in the AFL. It is sung to the tune of "Goodbye, Dolly Gray", originally a song written in connection with the Spanish–American War, then a popular Boer War and First World War anthem. It is the only AFL team song to reference the barracker, an Australian rules football term for fan.
The current version of the song played at the ground during game day was recorded in 1972 by the Fable Singers. The lyrics are as follows:
Good old Collingwood forever,
They know how to play the game.
Side by side, they stick together,
To uphold the Magpies name.
See, the barrackers are shouting,
As all barrackers should.
Oh, the premiership's a cakewalk,
For the good old Collingwood.
Rivalries
Carlton
Carlton is considered to be the club's most bitter arch-rival (for full details see Carlton–Collingwood AFL rivalry), with Richmond, Essendon and more recently Brisbane close behind. Collingwood's two opponents in the themed Rivalry Rounds staged to date have been Carlton (2005–2006, 2009) and Richmond (2007–2008).
Brisbane
Angst between supporters of Collingwood and Brisbane had been caused by plenty of history between the two clubs, despite the Brisbane Lions having a relatively short existence as a merged club. One of its merger clubs, Fitzroy topped the VFL/AFL premiership tally during the early existence of the VFL competition, alongside Collingwood. There was also animosity between the Brisbane Bears and the Magpies after the Bears' number one draft pick Nathan Buckley famously defected to Collingwood after one season on the Bears list. The rivalry between the two clubs peaked in the early 2000s, as the clubs played off in two consecutive Grand Finals in 2002 and 2003, with the Lions emerging victors on both occasions.
Melbourne
The rivalry between Collingwood and Melbourne was at its peak between 1955 and 1964, when the two played off in the grand final on five occasions. This included the 1958 Grand Final where Collingwood's victory prevented Melbourne from equalling Collingwood's record of four premierships in succession (1927–1930). The old rivalry with Melbourne has faded in recent decades due to Melbourne not enjoying the same level of on-field success, however, it remains strong and is an annual scheduled fixture on the Queens Birthday public holiday.
Essendon
Collingwood's rivalry with Essendon has become more significant since 1995, when the first Anzac Day clash took place. After the 2021 match, Collingwood have won this contest 15 times and Essendon 11 times, with the first match being drawn.
Geelong
Games between Collingwood and Geelong have become highly anticipated since 2007. In Round 15 Geelong beat Collingwood by 16 points in a high-quality match. In the Preliminary final Collingwood surprised many when they came within 5 points of the eventual premiers. In 2008 Collingwood thrashed Geelong by 86 points—20.14 (134)- 7.6 (48) causing Geelong's only loss of the 2008 home-and-away season. In 2009, the sides again met in the preliminary final, but despite high hopes the Cats, who would again win the premiership, won by 73 points in front of another massive crowd of 87,258. In 2010, the two sides emerged as the favourites for the flag and twice met in front of blockbuster crowds at the MCG when they were placed 1st and 2nd on the ladder—with the results evenly split. They again met in a Preliminary final, this time a resounding win to Collingwood by 41 points. In 2011, both teams were undefeated going into their round eight 'blockbuster' at the 'G. Geelong won by three points, after a controversial advantage was not paid to Magpie Scott Pendlebury in the dying minutes. Pendlebury kicked a goal and would have put the Pies in front, but the free kick was contentiously called back and Geelong managed to whisk the ball away. In the round 24 match, Geelong thumped the Magpies by a record margin of 96 points, which was also Collingwood's biggest ever loss at the MCG. The 2011 Grand Final against the Cats concluded with a 38-point loss for the Pies.
Supporters
Collingwood is a working-class suburb and the Collingwood Football Club supporter base traditionally came from the working class (though its supporter base today goes far beyond). Many of the club's supporters who regularly attend games still come from the working class or from lower socio-economic groups, leading to jokes from supporters of other clubs which typically stereotype their Collingwood counterparts as poor, crude and ignorant.
Collingwood is traditionally reviled by non-Collingwood supporters ("You either love 'em or you hate 'em"). The dislike of the club by outsiders is said to have originated during the 1920s and 1930s, a period of great success for the club which drew the envy and resentment of other clubs. In this period, Collingwood was also perceived as a Catholic and Irish club, at a time when these groups were looked down upon by the rest of Australian society and subjected to a considerable degree of social exclusion.
Until recent years, racial vilification of opposition players in VFL/AFL games was not uncommon. According to a 2001 study, Collingwoods old home ground of Victoria Park had a reputation as one of the worst venues for such vilification, though it has also been said that the problem was similar at all grounds. Collingwood has however been involved in several high-profile incidents of this type, such as those involving indigenous players Nicky Winmar in 1993 and Adam Goodes in 2013. Michael Longs accusation of racial vilification against Collingwood ruckman Damian Monkhorst in 1995 also led directly to the establishment of the AFL's racial vilification regulations. In support of more inclusive sporting cultures, in 2010 the Australian fashion designer Shanaaz Copeland developed a Collingwood-inspired hijab for Muslim women. (See also: The "Do Better" Report)
Collingwood’s cultural reach on the Australian sporting landscape is far-reaching, as evident by attendance figures, memberships, TV ratings, and, more recently, the emergence of digital media such as the Pie Hard podcast.
Corporate
Membership
In 2011, Collingwood reached 70,000 members for the first time, creating a new AFL record, beating their own previous record of 58,249 set in 2010.
The club's extensive membership base tends to be a large crowd-pulling power, which has caused the AFL to be accused of favouring Collingwood when scheduling to maximise the league's attendance figures. However, the AFL states that this is due to other clubs requesting home games at the MCG against Collingwood.
Off-field
Collingwood was one of the last clubs to abandon its traditional stadium, the famous inner-city Victoria Park. Collingwood now plays home games at the MCG. It now also has its headquarters situated in the former Glasshouse Entertainment Centre. Due to a sponsorship deal, this facility is known as 'The Holden Centre'.
Collingwood continues to be financially viable through the loyal support of its huge following and numerous sponsors. After finishing 2nd in 2002 and 2003 the team fell to 13th and 15th (out of 16) in 2004 and 2005 respectively. This trend has plagued the club since the glory days of pre-World War II VFL football. Since 1958, the club has won only two VFL/AFL Premiership (the inaugural AFL Premiership in 1990, and in 2010). Despite this, the club still has won more individual games, more finals and made more grand final appearances than any other club.
On 9 March 2007, former Collingwood and Fitzroy defender Gary Pert was appointed the Magpies' CEO, seven weeks after Greg Swann departed for Carlton. In accepting the key Magpie post, Pert quit as a club director and as managing director of Channel 9 in Melbourne. In a press conference, it was stated that Collingwood has budgeted to turn over about $50 million this year. McGuire hopes the new administration will soon double that figure. "A finance administration review has come up with how we are going to turn Collingwood in to its next phase of its life", McGuire said. "What do we do to make ourselves go from a $45 million a year turnover business to a $100 million turnover business? "They sound like big figures but in 1999 we turned over $13 million, so that is where we are heading as a football club."
The club made an operating profit of $5.23 million for the 2013 season, revenue increased from $2.6 million to more than $75 million.
On 24 July 2017, Pert resigned from his position as CEO of the club, with Peter Murphy replacing him as an interim CEO.
Sponsorship
The Collingwood guernsey is the most valuable sports sponsorship in Australia. Collingwood has different guernsey sponsors for home and away matches, generating an estimated $6.3 million worth of media exposure for the primary sponsor and $5.7 million for the secondary sponsor. These sponsorships are ranked first and second in Australia. High-profile sponsors include Emirates, Holden, CGU Insurance, and Westpac.
Honours
Honour board
Head-to-head results
Played: 2,598 Won: 1,556 Drawn:28 Lost: 986
(Last updated – End of 2020 AFL season)
Team of the Century
Collingwood announced its team of the century on 14 June 1997, celebrating 100 years since the beginning of the VFL. Gavin Brown was added as the fourth interchange player in 2002, as, when the team was named in 1997, only three interchange players were permitted on a team.
Captains
This list comprises every captain of the club. This list does not include deputy captains filling in due to an injury to the named captain, but does include captains named after a player retires or steps down during the season.
Bill Strickland 1897
Bill Proudfoot 1898–99, 1901
Dick Condon 1899–1900
Lardie Tulloch 1902–04
Charlie Pannam 1905
Alf Dummett 1906
Arthur Leach 1906–08
Eddie Drohan 1908
Robert Nash 1908–09
George Angus 1910–11
Jock McHale 1912–13
Dan Minogue 1914–16
Percy Wilson 1917–18
Con McCarthy 1919
Dick Lee 1920–21
Tom Drummond 1922
Harry Curtis 1923
Charlie Tyson 1924–26
Syd Coventry 1927–34
Harry Collier 1935–39
Jack Regan 1940–41, 1943
Phonse Kyne 1942, 1946–49
Pat Fricker 1944
Alby Pannam 1945
Gordon Hocking 1950–51
Lou Richards 1952–55
Neil Mann 1955–56
Bill Twomey Jr. 1957
Frank Tuck 1958–59
Murray Weideman 1960–63
Ray Gabelich 1964–65
John Henderson 1965
Des Tuddenham 1966–69, 1976
Terry Waters 1970–71
Wayne Richardson 1971–75
Max Richardson 1977
Len Thompson 1978
Ray Shaw 1979–80
Peter Moore 1981–82
Mark Williams 1983–86
Tony Shaw 1987–93
Gavin Brown 1994–98
Nathan Buckley 1999–2007
Scott Burns 2008
Nick Maxwell 2009–2013
Scott Pendlebury 2014–
Presidents
There have been twelve presidents of the Collingwood Football Club. The first and founding president of Collingwood was former Collingwood Mayor and Victorian MP William Beazley. Beazley was president of Collingwood from the founding of the club in 1892 until 1911. The second president of Collingwood was Alfred Cross. However, Cross was only president for a brief period of time. Third was former Fitzroy and Collingwood player Jim Sharp. Sharp was president for ten years (1913–1923). The fourth president of Collingwood was another former player, Harry Curtis. Curtis currently is the longest serving president of Collingwood. Curtis served as president for twenty-six years. Another former player of Collingwood, Syd Coventry was the fifth president for Collingwood, serving twelve years between 1950 and 1962.
Tom Sherrin was the sixth president of Collingwood, serving from 1963 to 1974. Ern Clarke, president for one year, was the seventh president. John Hickey, Ranald Macdonald and Allan MacAlister all served as president during 1977 through to 1995. Eleventh president and former player, Kevin Rose, was the second most recent president of Collingwood. The twelfth, and second-longest serving president of Collingwood, is radio and television presenter, commentator and journalist Eddie McGuire. McGuire was president of Collingwood between 1998 and 2021. Club board members Mark Korda and Peter Murphy were interim co-presidents, following McGuire's tenure. In April 2021, Korda was appointed the thirteenth president of Collingwood.
Current playing squad
Reserves team
The Collingwood Reserves are the reserves team of the club. The latest iteration of the Collingwood Reserves was created in 2008, and are yet to win a Victorian Football League (VFL) premiership.
History
The VFL/AFL operated a reserves competition from 1919 to 1991, and a de facto AFL reserves competition was run by the Victorian State Football League from 1992 to 1999. Collingwood fielded a reserves team in both of these competitions, allowing players who were not selected for the senior team to play for Collingwood in the lower grade. The team won seven reserves premierships during this period, including four in the first seven years between 1919 and 1925, but only three thereafter. After the AFL reserves competition was disbanded at the end of 1999, the club fielded its reserves team in the Victorian Football League during the 2000 season.
In 2001, Collingwood reserves team was dissolved and the club entered into an affiliation with the VFL's Williamstown Football Club, such that Williamstown served as a feeder team and reserves players for Collingwood played senior football for Williamstown. Williamstown won one VFL premiership during this time, in 2003.
Collingwood ended its affiliation with Williamstown after the 2007 season. The reserves team was re-established, and has competed in the VFL since 2008. Collingwood's standalone reserves team's best VFL result to date was a preliminary final appearance in the 2016 VFL season, in which it lost to eventual premiers Footscray by 119 points.
The reserves team currently splits home games between Olympic Park Oval and Victoria Park, although they do occasionally play at the MCG as a curtain raiser to Collingwood home matches, and uses the AFL team's clash guernsey as its primary guernsey. The Collingwood VFL team is composed of both reserves players from the club's primary and rookie AFL lists, and a separately maintained list of players eligible only for VFL matches.
Coaches
Note: Garry Hocking was appointed coach for the 2020 season, which was abandoned due to the COVID-19 pandemic.
Captains
Honours
Season summaries
Sources: Collingwood Football Club VFL Honour Roll, Collingwood Reserves Honour Roll 1919–2020, VFL Stats
Women's teams
History
In April 2016, the club launched a bid to enter a team in the inaugural AFL Women's season in 2017. Meg Hutchins was appointed Women's Football Operations Manager some weeks prior, and given the responsibility of crafting the bid.
The club was granted a license in June 2016, becoming one of eight teams to compete in the league's first season.
In addition to her role off-field, Hutchins would become one of the club's first players, along with marquees Moana Hope and Emma King. Collingwood selected a further 19 players in October's inaugural draft as well as three non-drafted players and two first time footballing rookies. Dandenong Stingrays assistant and Victorian Metro Youth Girls head coach Wayne Siekman was appointed the team's inaugural head coach in July 2016.
The club began fielding its own team in the revamped VFL Women's league from the start of the 2018 season. Many of the club's AFLW athletes play for the VFLW team, though the majority of the team is made up of players who haven't been drafted to an AFLW club. The VFL Women's competition runs from May to September (after the AFL Women's season has concluded) and Collingwood achieved success quickly in the league, claiming their first VFLW premiership in 2019.
The women's teams are based at the club's training and administration at Olympic Park, though often share matches between the venue and the club's spiritual home Victoria Park.
AFL Women's squad
Season summaries
AFL Women's
^ Denotes the ladder was split into two conferences. Figure refers to the club's overall finishing in the home-and-away season.
VFL Women's
Sources: Club historical data and VFLW Stats
Individual awards
Best and Fairest
Brownlow Medal winners
Syd Coventry (1927)
Albert Collier (1929)
Harry Collier (1930 tied)
Marcus Whelan (1939)
Des Fothergill (1940 tied)
Len Thompson (1972)
Peter Moore (1979)
Nathan Buckley (2003 tied)
Dane Swan (2011)
Leigh Matthews Trophy winners
Darren Millane (1990)
Dane Swan (2010)
Coleman Medal winners
Instituted in 1981, retrospective awards were dated back to 1955; prior to that, the League awarded the Leading Goalkicker Medal
Ian Brewer (1958)
Peter McKenna (1972, 1973)
Brian Taylor (1986)
Leading Goalkicker Medal winners
Archie Smith 1898
Teddy Lockwood 1900 (tied), 1903
Charlie Pannam 1905
Dick Lee 1907, 1908, 1909, 1914, 1916, 1917, 1919
Gordon Coventry 1926, 1927, 1928, 1929, 1930, 1933
Ron Todd 1938, 1939
Des Fothergill 1946
Norm Smith Medal winners
Tony Shaw (1990)
Nathan Buckley (2002)
Scott Pendlebury (2010)
E. J. Whitten Medalists
Gavin Brown (1989, 1997)
Mark of the Year winners
Alan Atkinson (1973)
Billy Picken (1974)
Billy Picken (1976)
Peter Daicos (1980)
Denis Banks (1984)
Chris Tarrant (2003)
Andrew Krakouer (2011)
Jamie Elliott (2013)
Goal of the Year winners
Phil Manassa (1977)
Peter Daicos (1991)
Mick McGuane (1994)
Leon Davis (2008)
Josh Daicos (2020)
Anzac Day Medal winners
Saverio Rocca (1995, 1998) ^
Scott Russell (1996) ^
Damien Monkhorst (1997) ^
Chris Tarrant (2001)
Mark McGough (2002)
Ben Johnson (2006)
Heath Shaw (2007)
Paul Medhurst (2008)
Scott Pendlebury (2010, 2011, 2019)
Dane Swan (2012, 2014)
Paul Seedsman (2015)
Steele Sidebottom (2016)
Adam Treloar (2018)
^ Awarded retrospectively in 2011
Neale Daniher Trophy winners
Travis Cloke (2015)
Mason Cox (2018)
Adam Treloar (2019)
Bob Rose-Charlie Sutton Medal winners
Ben Johnson (2008)
Dane Swan (2009)
Scott Pendlebury (2010, 2012, 2017)
Heath Shaw (2011, 2013)
Tom Phillips (2018)
Richard Pratt Medal winners
Dane Swan (2013)
Tom Langdon (2014)
Scott Pendlebury (2015)
Steele Sidebottom (2018)
Jason McCartney Medal winners
Anthony Rocca (2003)
Ben Johnson (2004)
Chris Tarrant (2006)
James Clement (2007)
Tarkyn Lockyer (2009)
Scott Pendlebury (2013)
Not awarded since 2013
All Australian Team
Des Healey (1953)
Bob Rose (1953)
Terry Waters (1969)
Ricky Watt (1969)
Peter McKenna (1972)
Len Thompson (1972)
Peter Moore (1979)
Michael Richardson (1983)
Geoff Raines (1985)
Tony Francis (1991)
Gavin Brown (1991, 1994)
Mick McGuane (1992)
Nathan Buckley (1996, 1997, 1998, 1999, 2001, 2003)
Chris Tarrant (2003)
James Clement (2004, 2005)
Alan Didak (2006, 2010)
Paul Medhurst (2008)
Dane Swan (2009, 2010, 2011, 2012, 2013)
Nick Maxwell (2009)
Leon Davis (2009, 2011)
Scott Pendlebury (2010, 2011, 2012, 2013, 2014, 2019)
Harry O'Brien (2010)
Dale Thomas (2011)
Ben Reid (2011)
Travis Cloke (2011, 2013)
Dayne Beams (2012)
Brodie Grundy (2018, 2019)
Steele Sidebottom (2018)
Adam Treloar (2019)
Darcy Moore (2020)
Taylor Adams (2020)
International rules representatives
Gavin Brown (1990)
Nathan Buckley (1998), (1999 – captain)
James Clement (2002)
Alan Didak (2004)
Scott Pendlebury (2008), (2017)
Dale Thomas (2008)
Dane Swan (2010)
Tyson Goldsack (2010)
Michael Tuck Medal winners
Heath Shaw (2011)
Jim Stynes Medal winners
Dane Swan (2010)
Match records
Highest score: R17, 1980 – Collingwood 32.19 (211) v St Kilda 16.11 (107) – Waverley Park
Lowest score: R6, 1897 (VP) – Collingwood 0.8 (8) v South Melbourne 2.15 (27) – Victoria Park (VP)
Lowest score since 1919: Grand Final, 1960 – 2.2 (14) v Melbourne 8.14 (62) – Melbourne Cricket Ground (MCG)
Highest losing score: R16, 1937 – Collingwood 21.16 (142) v Melbourne 22.21 (153) – VP
Lowest winning score: R9, 1899 (VP) – Collingwood 3.3 (21) v Melbourne 1.7 (13) – VP
Lowest winning score since 1919: Grand Final, 1927 – 2.13 (25) v Richmond 1.7 (13) – MCG
Biggest winning margin: 178 points; R4, 1979 – Collingwood 31.21 (207) v St Kilda 3.11 (29) – VP
Biggest losing margin: 138 points; R3, 1942 – Collingwood 5.7 (37) v Richmond 25.25 (175) – Punt Road Oval
Record attendance (home and away game): R10, 1958 – 99,346 v Melbourne – MCG
Record attendance (finals match): Grand Final, 1970 – 121,696 v Carlton – MCG
Records set by players
Most matches: Scott Pendlebury – 326 (2006–)
Most consecutive matches: Jock McHale – 191 (1906–1917) – VFL record until 1943
Most goals kicked in a match: Gordon Coventry – 17 goals 4 behinds (R12, 1930, VP) – VFL record until 1947
Most Best & Fairests: Nathan Buckley – 6 (1994, 1996, 1998, 1999, 2000, 2003)
Most matches as coach: Jock McHale – 714 (1912–1949) – VFL/AFL record until 2015 (Remains a record for the most matches as coach at one club.)
Most matches as captain/acting captain: Scott Pendlebury – 162 (2014–)
Most goals in a season: Peter McKenna – 143 (1970)
Most career goals: Gordon Coventry – 1299 (1920–1937) – VFL/AFL record until 1999 (Remains a record for the most career goals at one club.)
In popular culture
David Williamson's 1977 stage play, The Club, was inspired by the backroom dealings and antics of the Collingwood Football Club; although Collingwood is never mentioned by name. The 1980 film version of the play – directed by Bruce Beresford and starring John Howard, Jack Thompson, Graham Kennedy and Frank Wilson – is set at Collingwood and featured Collingwood players in speaking and non-speaking roles. The film was almost entirely shot on location at Victoria Park, both inside and on the actual oval.
Judd Apatow's 2009 film, Funny People, starring Adam Sandler and Seth Rogen, featured a scene with Australian actor Eric Bana trying to explain the rules of Australian rules football. During this scene Bana's character, a St Kilda supporter, voices his dislike for Collingwood while watching a televised game.
Adam Elliot's 2009 clay-animated film, Mary and Max, features a scene with a school-yard bully, named Bernie Clifford, who wears a 1970s VFL style Collingwood guernsey. A Collingwood garden gnome can also be seen in the film.
In the 2010 independent Australian film Joffa: The Movie, Joffa Corfe and Shane McRae star as a couple of knockabout handymen with a passion for the Collingwood Football Club.
John Brack's 1953 painting Three of the Players depicts three Collingwood players. The players are thought by some to be Lou Richards, Jack Regan and Phonse Kyne.
See also
History of the Collingwood Football Club
List of Collingwood players
List of Collingwood Football Club coaches
Sport in Australia
Sport in Victoria
Footnotes
Notes
References
Bibliography
Victorian Government Hansard of November 1990, pp.2208–2218: Victorian Legislative Assembly's debate on the Collingwood (Victoria Park) Land Bill on 21 November 1990: features an informative interchange between Murray Weideman's older brother, Graeme Weideman, and former South Melbourne footballer, Bill McGrath, both of whom were MLAs at the time.
External links
Australian rules football clubs established in 1892
Australian Football League clubs
Australian rules football clubs in Melbourne
Former Victorian Football League clubs
1892 establishments in Australia
AFL Women's clubs |
null | null | The Birth of Venus | eng_Latn | The Birth of Venus ( ) is a painting by the Italian artist Sandro Botticelli, probably executed in the mid 1480s. It depicts the goddess Venus arriving at the shore after her birth, when she had emerged from the sea fully-grown (called Venus Anadyomene and often depicted in art). The painting is in the Uffizi Gallery in Florence, Italy.
Although the two are not a pair, the painting is inevitably discussed with Botticelli's other very large mythological painting, the Primavera, also in the Uffizi. They are among the most famous paintings in the world, and icons of the Italian Renaissance; of the two, the Birth is better known than the Primavera. As depictions of subjects from classical mythology on a very large scale they were virtually unprecedented in Western art since classical antiquity, as was the size and prominence of a nude female figure in the Birth. It used to be thought that they were both commissioned by the same member of the Medici family, but this is now uncertain.
They have been endlessly analysed by art historians, with the main themes being: the emulation of ancient painters and the context of wedding celebrations (generally agreed), the influence of Renaissance Neo-Platonism (somewhat controversial), and the identity of the commissioners (not agreed). Most art historians agree, however, that the Birth does not require complex analysis to decode its meaning, in the way that the Primavera probably does. While there are subtleties in the painting, its main meaning is a straightforward, if individual, treatment of a traditional scene from Greek mythology, and its appeal is sensory and very accessible, hence its enormous popularity.
Description and subject
In the centre the newly-born goddess Venus stands nude in a giant scallop shell. The size of the shell is purely imaginary, and is also found in classical depictions of the subject. At the left the wind god Zephyr blows at her, with the wind shown by lines radiating from his mouth. He is in the air, and carries a young female, who is also blowing, but less forcefully. Both have wings. Vasari was probably correct in identifying her as "Aura", personification of a lighter breeze. Their joint efforts are blowing Venus towards the shore, and blowing the hair and clothes of the other figures to the right.
At the right a female figure who may be floating slightly above the ground holds out a rich cloak or dress to cover Venus when she reaches the shore, as she is about to do. She is one of the three Horae or Hours, Greek minor goddesses of the seasons and of other divisions of time, and attendants of Venus. The floral decoration of her dress suggests she is the Hora of Spring.
Alternative identifications for the two secondary female figures involve those also found in the Primavera; the nymph held by Zephyr may be Chloris, a flower nymph he married in some versions of her story, and the figure on land may be Flora. Flora is generally the Roman equivalent of the Greek Chloris; in the Primavera Chloris is transformed into the figure of Flora next to her, following Ovid's Fasti, but it is hard to see that such a transformation is envisaged here. However, the roses blown along with the two flying figures would be appropriate for Chloris.
The subject is not strictly the "Birth of Venus", a title given to the painting only in the nineteenth century (though given as the subject by Vasari), but the next scene in her story, where she arrives on land, blown by the wind. The land probably represents either Cythera or Cyprus, both Mediterranean islands regarded by the Greeks as territories of Venus.
Technical
The painting is large, but slightly smaller than the Primavera, and where that is a panel painting, this is on the cheaper support of canvas. Canvas was increasing in popularity, perhaps especially for secular paintings for country villas, which were decorated more simply, cheaply and cheerfully than those for city palazzi, being designed for pleasure more than ostentatious entertainment.
The painting is on two pieces of canvas, sewn together before starting, with a gesso ground tinted blue. There are differences to Botticelli's usual technique, working on panel supports, such as the lack of a green first layer under the flesh areas. There are a number of pentimenti revealed by modern scientific testing. The Hora originally had "low classical sandals", and the collar on the mantle she holds out is an afterthought. The hair of Venus and the flying couple was changed. There is heavy use of gold as a pigment for highlights, on hair, wings, textiles, the shell and the landscape. This was all apparently applied after the painting was framed. It was finished with a "cool gray varnish", probably using egg yolk.
As in the Primavera, the green pigment – used for the wings of Zephyr, Zephyr's companion, and the leaves of the orange trees on the land – has darkened considerably with exposure to light over time, somewhat distorting the intended balance of colours. Parts of some leaves at the top right corner, normally covered by the frame, have been less affected. The blues of the sea and sky have also lost their brightness.
Style
Although the pose of Venus is classical in some respects, and borrows the position of the hands from the Venus Pudica type in Greco-Roman sculptures (see section below), the overall treatment of the figure, standing off-centre with a curved body of long flowing lines, is in many respects from Gothic art. Kenneth Clark wrote: "Her differences from antique form are not physiological, but rhythmic and structural. Her whole body follows the curve of a Gothic ivory. It is entirely without that quality so much prized in classical art, known as aplomb; that is to say, the weight of the body is not distributed evenly either side of a central plumb line. .... She is not standing but floating. ... Her shoulders, for example, instead of forming a sort of architrave to her torso, as in the antique nude, run down into her arms in the same unbroken stream of movement as her floating hair."
Venus' body is anatomically improbable, with elongated neck and torso. Her pose is impossible: although she stands in a classical contrapposto stance, her weight is shifted too far over the left leg for the pose to be held. The proportions and poses of the winds to the left do not quite make sense, and none of the figures cast shadows. The painting depicts the world of the imagination rather than being very concerned with realistic depiction.
Ignoring the size and positioning of the wings and limbs of the flying pair on the left, which bother some other critics, Kenneth Clark calls them:
...perhaps the most beautiful example of ecstatic movement in the whole of painting. ... the suspension of our reason is achieved by the intricate rhythms of the drapery which sweep and flow irresistibly around the nude figures. Their bodies, by an endless intricacy of embrace, sustain the current of movement, which finally flickers down their legs and is dispersed like an electric charge.
Botticelli's art was never fully committed to naturalism; in comparison to his contemporary Domenico Ghirlandaio, Botticelli seldom gave weight and volume to his figures and rarely used a deep perspectival space. Botticelli never painted landscape backgrounds with great detail or realism, but this is especially the case here. The laurel trees and the grass below them are green with gold highlights, most of the waves regular patterns, and the landscape seems out of scale with the figures. The clumps of bulrushes in the left foreground are out of place here, as they come from a freshwater species.
Dating and history
It has long been suggested that Botticelli was commissioned to paint the work by the Medici family of Florence, perhaps by Lorenzo di Pierfrancesco de' Medici (1463–1503) a major patron of Botticelli, under the influence of his cousin Lorenzo de' Medici, "il Magnifico". This was first suggested by Herbert Horne in his monograph of 1908, the first major modern work on Botticelli, and long followed by most writers, but more recently has been widely doubted, though it is still accepted by some. Various interpretations of the painting rely on this origin for its meaning. Although relations were perhaps always rather tense between the Magnifico and his young cousins and wards, Lorenzo di Pierfrancesco and his brother Giovanni di Pierfrancesco de' Medici, it may have been politic to commission a work that glorified the older Lorenzo, as some interpretations have it. There may be a deliberate ambiguity as to which Lorenzo was intended to be evoked. In later years hostility between the two branches of the family became overt.
Horne believed that the painting was commissioned soon after the purchase in 1477 of the Villa di Castello, a country house outside Florence, by Lorenzo and Giovanni, to decorate their new house, which they were rebuilding. This was the year after their father died at the age of 46, leaving the young boys wards of their cousin Lorenzo il Magnifico, of the senior branch of the Medici family and de facto ruler of Florence. There is no record of the original commission, and the painting is first mentioned by Vasari, who saw it, together with the Primavera, at Castello, some time before the first edition of his Lives in 1550, probably by 1530–40. In 1550 Vasari was himself painting in the villa, but he very possibly visited it before that. But in 1975 it emerged that, unlike the Primavera, the Birth is not in the inventory, apparently complete, made in 1499 of the works of art belonging to Lorenzo di Pierfrancesco's branch of the family. Ronald Lightbown concludes that it only came to be owned by the Medici after that. The inventory was only published in 1975, and made many previous assumptions invalid.
Horne dated the work at some point after the purchase of the villa in 1477 and before Botticelli's departure for Rome to join the painting of the Sistine Chapel in 1481. Recent scholars prefer a date of around 1484–86 on grounds of the work's place in the development of Botticelli's style. The Primavera is now usually dated earlier, after Botticelli's return from Rome in 1482 and perhaps around the time of Lorenzo di Pierfrancesco's wedding in July 1482, but by some still before Botticelli's departure.
Whenever the two paintings were united at Castello, they have remained together ever since. They stayed in Castello until 1815, when they were transferred to the Uffizi. For some years until 1919 they were kept in the Galleria dell'Accademia, another government museum in Florence.
Interpretations
Although there are ancient and modern texts that are relevant, no single text provides the precise imagery of the painting, which has led scholars to propose many sources and interpretations. Many art historians who specialize in the Italian Renaissance have found Neoplatonic interpretations, of which two different versions have been articulated by Edgar Wind and Ernst Gombrich, to be the key to understanding the painting. Botticelli represented the Neoplatonic idea of divine love in the form of a nude Venus.
For Plato – and so for the members of the Florentine Platonic Academy – Venus had two aspects: she was an earthly goddess who
aroused humans to physical love or she was a heavenly goddess who inspired intellectual love in them. Plato further argued that contemplation of physical beauty allowed the mind to better understand spiritual beauty. So, looking at Venus, the most beautiful of goddesses, might at first raise a physical response in viewers which then lifted their minds towards the godly. A Neoplatonic reading of Botticelli's Birth of Venus suggests that 15th-century viewers would have looked at the painting and felt their minds lifted to the realm of divine love.
The composition, with a central nude figure, and one to the side with an arm raised above the head of the first, and winged beings in attendance, would have reminded its Renaissance viewers of the traditional iconography of the Baptism of Christ, marking the start of his ministry on earth. In a similar way, the scene shows here marks the start of Venus's ministry of love, whether in a simple sense, or the expanded meaning of Renaissance Neoplatonism.
More recently, questions have arisen about Neoplatonism as the dominant intellectual system of late 15th-century Florence, and scholars have indicated that there might be other ways to interpret Botticelli's mythological paintings. In particular, both Primavera and Birth of Venus have been seen as wedding paintings that suggest appropriate behaviors for brides and grooms.
The laurel trees at right and laurel wreath worn by the Hora are punning references to the name "Lorenzo", though it is uncertain whether Lorenzo il Magnifico, the effective ruler of Florence, or his young cousin Lorenzo di Pierfrancesco is meant. In the same way the flowers in the air around Zephyr and on the textiles worn and carried by the Hora evoke the name of Florence.
Literary sources
The closest precedent for the scene is generally agreed to be in one of the early ancient Greek Homeric Hymns, published in Florence in 1488 by the Greek refugee Demetrios Chalkokondyles:
Of august gold-wreathed and beautiful
Aphrodite I shall sing to whose domain
belong the battlements of all sea-loved
Cyprus where, blown by the moist breath
of Zephyros, she was carried over the
waves of the resounding sea on soft foam.
The gold-filleted Horae happily welcomed
her and clothed her with heavenly raiment.
This poem was probably already known to Botticelli's Florentine contemporary, and Lorenzo di Medici's court poet, Angelo Poliziano. The iconography of The Birth of Venus is similar to a description of a relief of the event in Poliziano's poem the Stanze per la giostra, commemorating a Medici joust in 1475, which may also have influenced Botticelli, although there are many differences. For example Poliziano talks of multiple Horae and zephyrs. Older writers, following Horne, posited that "his patron Lorenzo di Pierfrancesco asked him to paint a subject illustrating the lines", and that remains a possibility, though one difficult to maintain so confidently today. Another poem by Politian speaks of Zephyr causing flowers to bloom, and spreading their scent over the land, which probably explains the roses he blows along with him in the painting.
Ancient art
Having a large standing female nude as the central focus was unprecedented in post-classical Western painting, and certainly drew on the classical sculptures which were coming to light in this period, especially in Rome, where Botticelli had spent 1481–82 working on the walls of the Sistine Chapel. The pose of Botticelli's Venus follows the Venus Pudica ("Venus of Modesty") type from classical antiquity, where the hands are held to cover the breasts and groin; in classical art this is not associated with the new-born Venus Anadyomene. What became a famous example of this type is the Venus de' Medici, a marble sculpture that was in a Medici collection in Rome by 1559, which Botticelli may have had opportunity to study (the date it was found is unclear).
The painter and the humanist scholars who probably advised him would have recalled that Pliny the Elder had mentioned a lost masterpiece of the celebrated ancient Greek painter, Apelles, representing Venus Anadyomene (Venus Rising from the Sea). According to Pliny, Alexander the Great offered his mistress, Campaspe, as the model for the nude Venus and later, realizing that Apelles had fallen in love with the girl, gave her to the artist in a gesture of extreme magnanimity. Pliny went on to note that Apelles' painting of Pankaspe as Venus was later "dedicated by Augustus in the shrine of his father Caesar." Pliny also stated that "the lower part of the painting was damaged, and it was impossible to find anyone who could restore it. ... This picture decayed from age and rottenness, and Nero ... substituted for it another painting by the hand of Dorotheus".
Pliny also noted a second painting by Apelles of Venus "superior even to his earlier one," that had been begun by the artist but left unfinished. The Roman images in various media showing the new-born Venus in a giant shell may well be crude derivative versions of these paintings. Botticelli could not have seen the frescos unearthed later in Pompeii, but may well have seen small versions of the motif in terracotta or engraved gems. The "House of Venus" in Pompeii has a life-size fresco of Venus lying in the shell, also seen in other works; in most other images she stands with her hands on her hair, wringing the water from it, with or without a shell.
The two-dimensionality of this painting may be a deliberate attempt to evoke the style of ancient Greek vase painting or frescos on the walls of Etruscan tombs, the only types of ancient painting known to Botticelli.
Mack
Another interpretation of the Birth of Venus is provided by art historian and author, Charles R. Mack. This interpretation takes much that is generally agreed, but Mack goes on to explain the painting as an allegory extolling the virtues of Lorenzo de' Medici. This has not been adopted by Renaissance art historians in general, and it remains problematic, since it depends on the painting being commissioned by the Medici, yet the work is not documented in Medici hands until well into the following century.
Mack sees the scene as inspired by both the Homeric Hymn and the ancient paintings. But something more than a rediscovered Homeric hymn was likely in the mind of the Medici family member who commissioned this painting from Botticelli. Once again, Botticelli, in his version of the Birth of Venus, might be seen as completing the task begun by his ancient predecessor Apelles, even surpassing him. Giving added support to this interpretation of Botticelli as a born-again Apelles is the fact that that very claim was voiced in 1488 by Ugolino Verino in a poem entitled "On Giving Praise to the History of Florence."
While Botticelli might well have been celebrated as a revivified Apelles, his Birth of Venus also testified to the special nature of Florence's chief citizen, Lorenzo de' Medici. Although it now seems that the painting was executed for another member of the Medici family, it likely was intended to celebrate and flatter its head, Lorenzo de' Medici. Tradition associates the image of Venus in Botticelli's painting with the famous beauty Simonetta Cattaneo Vespucci, of whom popular legend claims both Lorenzo and his younger brother, Giuliano, were great admirers. Simonetta was possibly born in the Ligurian seaside town of Portovenere ('the port of Venus'). Thus, in Botticelli's interpretation, Pankaspe (the ancient living prototype of Simonetta), the mistress of Alexander the Great (the Laurentian predecessor), becomes the lovely model for the lost Venus executed by the famous Greek painter Apelles (reborn through the recreative talents of Botticelli), which ended up in Rome, installed by Emperor Augustus in the temple dedicated to Florence's supposed founder Julius Caesar.
In the case of Botticelli's Birth of Venus, the suggested references to Lorenzo, supported by other internal indicators such as the stand of laurel bushes at the right, would have been just the sort of thing erudite Florentine humanists would have appreciated. Accordingly, by overt implication, Lorenzo becomes the new Alexander the Great with an implied link to both Augustus, the first Roman emperor, and even to Florence's legendary founder, Caesar himself. Lorenzo, furthermore, is not only magnificent but, as was Alexander in Pliny's story, also magnanimous, as well. Ultimately, these readings of the Birth of Venus flatter not only the Medici and Botticelli but all of Florence, home to the worthy successors to some of the greatest figures of antiquity, both in governance and in the arts.
These essentially pagan readings of Botticelli's Birth of Venus should not exclude a more purely Christian one, which may be derived from the Neoplatonic reading of the painting indicated above. Viewed from a religious standpoint, the nudity of Venus suggests that of Eve before the Fall as well as the pure love of Paradise. Once landed, the goddess of love will don the earthly garb of mortal sin, an act that will lead to the New Eve – the Madonna whose purity is represented by the nude Venus. Once draped in earthly garments she becomes a personification of the Christian Church which offers a spiritual transport back to the pure love of eternal salvation. In this case the scallop shell upon which this image of Venus/Eve/Madonna/Church stands may be seen in its traditionally symbolic pilgrimage context. Furthermore, the broad expanse of sea serves as a reminder of the Virgin Mary's title stella maris, alluding both to the Madonna's name (Maria/maris) and to the heavenly body (Venus/stella). The sea brings forth Venus just as the Virgin gives birth to the ultimate symbol of love, Christ.
Rather than choosing one of the many interpretations offered for Botticelli's depiction of the Birth (Arrival?) of Venus it might be better to view it from a variety of perspectives. This layered approach—mythological, political, religious—was intended.
Derivative versions
Botticelli, or more likely his workshop, repeated the figure of Venus in another painting of about 1490. This life-sized work depicts a similar figure and pose, partially clad in a light blouse, and contrasted against a plain dark background. It is in the Galleria Sabauda in Turin. There is another such workshop Venus in Berlin, and very likely others were destroyed in the "Bonfire of the Vanities". Examples seem to have been exported to France and Germany, probably influencing Lucas Cranach the Elder among others.
More than a decade later, Botticelli adapted the figure of Venus for a nude personification of "Truth" in his Calumny of Apelles. Here one hand is raised, pointing to heaven for justification, and the figure's gaze also looks upwards; the whole effect is very different.
See also
100 Great Paintings
Notes
References
Clark, Kenneth, The Nude, A Study in Ideal Form, orig. 1949, various edns, page refs from Pelican edn of 1960
Dempsey, Charles, "Botticelli, Sandro", Grove Art Online, Oxford Art Online. Oxford University Press. Web. 15 May. 2017. subscription required.
"Ettlingers": Leopold Ettlinger with Helen S. Ettlinger, Botticelli, 1976, Thames and Hudson (World of Art),
Hartt, Frederick, History of Italian Renaissance Art, (2nd edn.)1987, Thames & Hudson (US Harry N Abrams),
Hemsoll, David, The Birth of Venus, University of Birmingham, 18 min introductory lecture, refs to mm:ss
Legouix, Susan, Botticelli, 115–118, 2004 (revd edn), Chaucer Press,
Lightbown, Ronald, Sandro Botticelli: Life and Work, 1989, Thames and Hudson
Mack, Charles R. (2002),"Botticelli's Venus: Antique Allusions and Medicean Propaganda," Explorations in Renaissance Culture, 28, 1 (Winter), 2002, 1–31.
Mack, Charles R. (2005), Looking at the Renaissance: Essays toward a Contextual Appreciation, Ann Arbor: University of Michigan Press, 2005
Wind, Edgar, Pagan Mysteries in the Renaissance, 1967 edn., Peregrine Books
External links
University of Birmingham: Dr David Hemsoll, The Birth of Venus – mini-lecture
ArtSleuth: The Birth of Venus – That Obscure Object of Desire
1486 paintings
Paintings by Sandro Botticelli in the Uffizi
Venus Anadyomenes
Nude art
Paintings depicting Greek myths
Paintings of Venus
Water in art
Seashells in art |
null | null | The Championships, Wimbledon | eng_Latn | The Championships, Wimbledon, commonly known simply as Wimbledon or The Championships, is the oldest tennis tournament in the world and is widely regarded as the most prestigious. It has been held at the All England Club in Wimbledon, London, since 1877 and is played on outdoor grass courts, with a retractable roof over two courts since 2019.
Wimbledon is one of the four Grand Slam tennis tournaments, the others being the Australian Open, the French Open and the US Open. Wimbledon is the only major still played on grass, which is the traditional tennis playing surface.
The tournament traditionally took place over two weeks in late June and early July, starting on the last Monday in June and culminating with the Ladies' and Gentlemen's Singles Finals, scheduled for the Saturday and Sunday at the end of the second week. However, changes to the tennis calendar in 2015 have seen the event moved back by a week to begin in early July. Five major events are held each year, with additional junior and invitational competitions also taking place.
Wimbledon traditions include a strict all-white dress code for competitors, and royal patronage. Strawberries and cream are traditionally consumed at the tournament. The tournament allows advertising around the courts only by the official sponsors, which are Rolex, which provides timekeeping technology during matches, IBM, Oppo, Slazenger, and Robinsons barley water.
In 2009, Wimbledon's Centre Court was fitted with a retractable roof to lessen the loss of playing time due to rain. A roof was operational over No. 1 Court from 2019, when a number of other improvements were made, including adding cushioned seating, a table and 10 independently operable cameras per court to capture the games.
Due to the COVID-19 pandemic, the 2020 Championships were cancelled, the first cancellation of the tournament since World War II. The 134th edition was staged from 28 June 2021 to 11 July 2021, following from the 2020 cancellation. The 135th edition is scheduled to take place between 27 June 2022 and 10 July 2022, with Sunday expected to be staged for the first time.
History
Beginning
The All England Lawn Tennis and Croquet Club is a private club founded on 23 July 1868, originally as "The All England Croquet Club". Its first ground was at Nursery Road off Worple Road, Wimbledon.
In 1876, lawn tennis, a game devised by Major Walter Clopton Wingfield a year or so earlier as an outdoor version of real tennis and originally given the name Sphairistikè, was added to the activities of the club. In spring 1877, the club was renamed "The All England Croquet and Lawn Tennis Club" and signalled its change of name by instituting the first Lawn Tennis Championship. A new code of laws, replacing the code administered by the Marylebone Cricket Club, was drawn up for the event. Today's rules are similar except for details such as the height of the net and posts and the distance of the service line from the net.
The inaugural 1877 Wimbledon Championship started on 9 July 1877 and the Gentlemen's Singles was the only event held. 22 men paid a guinea to enter the tournament, which was to be held over five days. The rain delayed it four more days and thus, on 19 July 1877, the final was played. Spencer Gore, an old Harrovian rackets player, defeated William Marshall 6–1, 6-2 and 6–4 in 48 minutes. Gore was presented with the silver challenge cup, valued at 25 guineas and donated by the sports magazine The Field, as well as a prize money of 12 guineas. About 200 spectators paid one shilling each to watch the final.
The lawns at the ground were arranged so that the principal court was in the middle with the others arranged around it, hence the title "Centre Court". The name was retained when the Club moved in 1922 to the present site in Church Road, although no longer a true description of its location. However, in 1980 four new courts were brought into commission on the north side of the ground, which meant the Centre Court was once more correctly described. The opening of the new No. 1 Court in 1997 emphasised the description.
By 1882, activity at the club was almost exclusively confined to lawn tennis and that year the word "croquet" was dropped from the title. However, for sentimental reasons it was restored in 1899.
In 1884, the club added the Ladies' Singles competition and the Gentlemen's Doubles was transferred from the Oxford University Lawn Tennis Club. Ladies' Doubles and Mixed Doubles events were added in 1913. The first black player to compete at Wimbledon was Bertrand Milbourne Clark, an amateur from Jamaica, in 1924.
Until 1922, the reigning champion had to play only in the final, against whomever had won through to challenge him/her. As with the other three Major or Grand Slam events, Wimbledon was contested by top-ranked amateur players; professional players were prohibited from participating. This changed with the advent of the open era in 1968. No British man won the singles event at Wimbledon between Fred Perry in 1936 and Andy Murray in 2013, while no British woman has won since Virginia Wade in 1977, although Annabel Croft and Laura Robson won the Girls' Championship in 1984 and 2008 respectively. The Championship was first televised in 1937.
Though properly called "The Championships, Wimbledon", depending on sources the event is also known as "The All England Lawn Tennis Championships", "The Wimbledon Championships" or simply "Wimbledon". From 1912 to 1924, the tournament was recognized by the International Lawn Tennis Federation as the "World Grass Court Championships".
In the period of 1915–1918, no tournament was organized due to World War I.
During World War II, the tournament was not held in the period 1940–1945. On 11 October 1940 one bomb hit a corner of the competitors’ stand of the Centre Court. The championships did go ahead in 1946 even though the damage meant that 1,200 seats were lost. The organisers were unable to repair the damaged section until 1947 and the Centre Court was fully restored and renovated for the 1949 edition.
In 1946 and 1947 Wimbledon was held before the French Championships and was thus the second Grand Slam tennis event of the year.
21st century
Wimbledon is widely considered the world's premier tennis tournament and the priority of the club is to maintain its leadership. To that end a long-term plan was unveiled in 1993, intended to improve the quality of the event for spectators, players, officials and neighbours. Stage one (1994–1997) of the plan was completed for the 1997 championships and involved building the new No. 1 Court in Aorangi Park, a broadcast centre, two extra grass courts and a tunnel under the hill linking Church Road and Somerset Road. Stage two (1997–2009) involved the removal of the old No. 1 Court complex to make way for the new Millennium Building, providing extensive facilities for players, press, officials and members, and the extension of the West Stand of the Centre Court with 728 extra seats. Stage three (2000–2011) was completed with the construction of an entrance building, club staff housing, museum, bank and ticket office.
A new retractable roof was built in time for the 2009 championships, marking the first time that rain did not stop play for a lengthy time on Centre Court. The Club tested the new roof at an event called A Centre Court Celebration on Sunday, 17 May 2009, which featured exhibition matches involving Andre Agassi, Steffi Graf, Kim Clijsters and Tim Henman. The first Championship match to take place under the roof was the completion of the fourth round women's singles match between Dinara Safina and Amélie Mauresmo. The first match to be played in its entirety under the new roof took place between Andy Murray and Stanislas Wawrinka on 29 June 2009. Murray was also involved in the match completed latest in the day at Wimbledon, which ended at 11:02pm in a victory over Marcos Baghdatis at Centre Court in the third round of the 2012 Championships. The 2012 Gentlemen's Singles Final on 8 July 2012, between Roger Federer and Murray, was the first singles final to be partially played under the roof, which was activated during the third set.
A new 4,000-seat No. 2 Court was built on the site of the old No. 13 Court in time for the 2009 Championships. A new 2,000-seat No. 3 Court was built on the site of the old No. 2 and No. 3 Courts.
On 1 August 2011, the All England Club transferred all of its assets relating to The Championships to a separate though wholly owned subsidiary, The All England Lawn Tennis Club (Championships) Limited, also known as AELTC. Since that time, the club's activities have been formally conducted separately from those of The Championships.
In 2012, the All England Club hosted the Summer Olympic Games and became the first Olympic grass court tournament since tennis was reintroduced as an Olympic sport and the first to be held at a Grand Slam venue in the Open era.
In April 2013, Wimbledon unveiled its 'Master Plan' a vision in which to improve the championships over the next 10–15 years. This was in large part due to other Grand Slam tournaments such as the French Open and Australian Open also announcing expansion and re-development plans. Aspects of the master plan included new player and media facilities, expansion of the No.1 court including a new retractable roof, new catering and hospitality areas, additional floor to the museum building, construction of an underground car park and new indoor courts and also a total reconfiguration of the site including the relocation of a number of practice, clay and championship courts.
Part of the plan also includes acquiring the land of the adjacent Wimbledon Park Golf club for £65 million, so that the tournament qualifying matches can be played on site.
On 19 October 2018, it was announced that a tie-break will be played if the score reaches 12–12 in the final set of any match; this will apply to all competitions including in qualifying, singles, and doubles. In a related statement, it was announced that starting at the 2019 Championships, quad wheelchair competitions would become a permanent event.
As a result of the COVID-19 global pandemic, the All England Club announced on 1 April 2020 that the entire grass-court season was to be cancelled as a public safety precaution, marking the first time a Wimbledon tournament would not be played since World War II. Club officials considered playing the tournament behind closed doors, but this was ruled out in part because at least 5,000 people–including ballboys, officials, coaches, maintenance, and security–would have still needed to be on site to hold a functioning tournament. Former player and current All England Club board member Tim Henman told the Tennis Channel of the US that the board had carefully considered holding a closed-door Wimbledon. However, the sheer number of people who still would have needed to be on site led the board to realise "that wasn't going to be a workable option". Prior to the start of the 2003 tournament, the club began paying an annual insurance premium of £1.61m ($2 million) to cover losses from cancellation of Wimbledon in the event of a worldwide pandemic as a result of the SARS outbreak; it would receive an insurance payment of £114 million ($141 million) for the 2020 cancellation on expected losses of around £250 million ($312 million).
Events
Wimbledon consists of five main events, four junior events and seven invitation events.
Main events
The five main events, and the number of players (or teams, in the case of doubles) are:
Gentlemen's Singles (128)
Ladies' Singles (128)
Gentlemen's Doubles (64)
Ladies' Doubles (64)
Mixed Doubles (48)
Junior events
The four junior events and the number of players or teams are:
Boys' Singles (64)
Girls' Singles (64)
Boys' Doubles (32)
Girls' Doubles (32)
No mixed doubles event is held at this level
Invitation events
The seven invitational events and the number of pairs are:
Gentlemen's Invitation Doubles (8 pairs Round Robin)
Ladies' Invitation Doubles (8 pairs Round Robin)
Senior Gentlemen's Invitation Doubles (8 pairs Round Robin)
Gentlemen's Wheelchair Singles
Ladies' Wheelchair Singles
Gentlemen's Wheelchair Doubles (4 pairs)
Ladies' Wheelchair Doubles (4 pairs)
Match formats
Matches in the Gentlemen's Singles and Doubles are best-of-five sets; all other events are best-of-three sets. Up to and including the 2018 tournament, a tiebreak game is played if the score reaches 6–all in any set except the fifth (in a five-set match) or the third (in a three-set match), in which case a two-game lead must be reached. Since 2019, a final set tiebreak game is played if the score in the final set reaches 12–all.
All events are single-elimination tournaments, except for the Gentlemen's, Senior Gentlemen's and the Ladies' Invitation Doubles, which are round-robin tournaments.
Up to 1921, the winners of the previous year's competition (except in the Ladies' Doubles and Mixed Doubles) were automatically granted byes into the final round (then known as the challenge round). This led to many winners retaining their titles in successive years, as they were able to rest while their opponent competed from the start of the competition. Since 1922, the prior year's champions were required to play all the rounds, like other tournament competitors.
Schedule
Each year the tournament begins on the last Monday in June or first Monday in July, two weeks after the Queen's Club Championships, which is one of the men's major warm-up tournaments, together with the Gerry Weber Open, which is held in Halle, Germany, during the same week. Other grass-court tournaments before Wimbledon are Eastbourne, Great Britain, and Rosmalen in the Netherlands, both combining mixed events. The other women's warm-up tournament for Wimbledon is Birmingham, also in Great Britain. The men's event which is outside Europe before Wimbledon is the Antalya open in Turkey. The only grass-court tournament scheduled after the Championships is the Hall of Fame Tennis Championships at Newport, Rhode Island, USA, which takes place the week after Wimbledon.
Wimbledon is scheduled for 14 days, beginning on a Monday and ending on a Sunday. Before 1982 it ended a day earlier, with the women's singles final on the Friday and the men's singles final on the Saturday. The five main events span both weeks, but the junior and invitational events are held mainly during the second week. Traditionally, unlike the other three tennis Grand Slams, there is no play on the "Middle Sunday", which is considered a rest day. However, rain has forced play on the Middle Sunday four times, in 1991, 1997, 2004 and 2016. On the first of these four occasions, Wimbledon staged a "People's Sunday", with unreserved seating and readily available, inexpensive tickets, allowing those with more limited means to sit on the show courts.
The second Monday at Wimbledon is often called "Manic Monday", because it is the busiest day with the last-16 matches for both men's and women's singles, where fans have a pick of watching on a single day, any of the best 32 players left; which is also unique in a Grand Slam singles competition.
From 2022, Wimbledon will make Middle Sunday a permanent fixture to the schedule. This would allow for more leeway for Round of 16 matches to be completed on schedule, as announced in their 2021 Spring Press Conference.
Since 2015, the championships have begun one week later than in previous years, extending the gap between the tournament and the French Open from two to three weeks. Additionally the Stuttgart Open men's tournament converted to a grass surface and was rescheduled from July to June, extending the grass court season.
Players and seeding
Both the men's and ladies' singles consist of 128 players. Players and doubles pairs are admitted to the main events on the basis of their international rankings, with 104 direct entries into the men's and 108 into the ladies' competitions. Both tournaments have 8 wild card entrants, with the remainder in each made up of qualifiers. Since the 2001 tournament, 32 players have been given seedings in the Gentlemen's and Ladies' singles, 16 teams in the doubles events. The system of seeding was introduced during the 1924 Wimbledon Championships. This was a simplified version allowing countries to nominate four players who were placed in different quarters of the draw. This system was replaced for the 1927 Wimbledon Championships and from then on players were seeded on merit. The first players to be seeded as no. 1 were René Lacoste and Helen Wills.
The Committee of Management decide which players receive wildcards. Usually, wild cards are players who have performed well during previous tournaments or would stimulate public interest in Wimbledon by participating. The only wild card to win the Gentlemen's Singles Championship was Goran Ivanišević in 2001. Players and pairs who neither have high enough rankings nor receive wild cards may participate in a qualifying tournament held one week before Wimbledon at the Bank of England Sports Ground in Roehampton. The singles qualifying competitions are three-round events. From 2019 singles qualification will increase to 128 players and no doubles qualification will occur. Previously the same-sex doubles competitions lasted for only two rounds. There is no qualifying tournament for Mixed Doubles. The furthest that any qualifier has progressed in a Singles tournament is the semi-final round: John McEnroe in 1977 (Gentlemen's Singles), Vladimir Voltchkov in 2000 (Gentlemen's Singles), and Alexandra Stevenson in 1999 (Ladies' Singles).
Players are admitted to the junior tournaments upon the recommendations of their national tennis associations, on their International Tennis Federation world rankings and, in the case of the singles events, on the basis of a qualifying competition. The Committee of Management determines which players may enter the four invitational events.
The Committee seeds the top players and pairs on the basis of their rankings, but it could change the seedings based on a player's previous grass court performance. Since 2002 a seeding committee has not been required for the Gentlemen's Singles following an agreement with the ATP, and since the 2021 tournament, the seeding has followed the same process as the ATP rankings. From 2002 to 2019, the top 32 players (according to the ATP rankings) were seeded according to a formula that more heavily weighted previous grass-court tournaments: ATP Entry System Position points + 100% points earned for all grass court tournaments in the past 12 months + 75% points earned for the best grass court tournament in the 12 months before that. A majority of the entrants are unseeded. Only two unseeded players have won the Gentlemen's Singles: Boris Becker in 1985 and Goran Ivanišević in 2001. In 1985 there were only 16 seeds and Becker was ranked 20th; Ivanišević was ranked 125th when he won as a Wild Card entrant, although he had previously been a finalist three times, and been ranked no. 2 in the world; his low ranking was due to having been hampered by a persistent shoulder injury for three years, which had only just cleared up. In 1996, the title was won by Richard Krajicek, who was originally unseeded (ranked 17th, and only 16 players were seeded) but was promoted to a seeded position (still with the number 17) when Thomas Muster withdrew before the tournament. No unseeded player has captured the Ladies' Singles title; the lowest seeded female champion was Venus Williams, who won in 2007 as the 23rd seed; Williams was returning from an injury that had prevented her playing in previous tournaments, giving her a lower ranking than she would normally have had. Unseeded pairs have won the doubles titles on numerous occasions; the 2005 Gentlemen's Doubles champions were not only unseeded, but also (for the first time ever) qualifiers.
Grounds
Since 2001, the courts used for Wimbledon have been sown with 100% perennial ryegrass. Prior to 2001 a combination of 70% ryegrass and 30% Creeping Red Fescue was used. The change was made to improve durability and strengthen the sward to better withstand the increasing wear of the modern game.
The main show courts, Centre Court and No. 1 Court, are normally used for only two weeks a year, during the Championships, but play can extend into a third week in exceptional circumstances. The remaining 17 courts are regularly used for other events hosted by the club. The show courts were in action for the second time in three months in 2012 as Wimbledon hosted the tennis events of the 2012 Olympic Games. One of the show courts is also used for home ties of the Great Britain teams in the Davis Cup on occasions.
Wimbledon is the only remaining Grand Slam event played on natural grass courts. At one time, all the Majors, except the French Open, were played on grass. The US Open abandoned grass in 1975 for green clay and the Australian Open did so in 1988 for hard courts; the US Open eventually adopted hard courts as well.
From 1877 until 1921, the Club's grounds were situated on four acres of meadowland in central Wimbledon between Worple Road and the railway line. In 1908, this venue hosted the tennis events for the 1908 Summer Olympic Games. As the attendance at the Championships grew, it became obvious before the First World War that the 8,000 ground capacity at Worple Road was inadequate, and so the Club started looking for a new site. It eventually settled on an area of land off Church Road, to the north of Wimbledon town centre, and moved to its new home in 1922. At the time the relocation was regarded as something of a financial gamble, costing as it did approximately £140,000. After the Club moved to the current site in Church Road, the old Worple Road ground then became the Wimbledon High School playing field, which it remains today.
The principal court at Church Road, Centre Court, was inaugurated in 1922. The new venue was substantially larger and was needed to meet the ever-growing public demand.
Due to the possibility of rain during Wimbledon, a retractable roof was installed prior to the 2009 Championship. It is designed to close/open fully in 20 minutes and will be closed primarily to protect play from inclement (and, if necessary, extremely hot) weather during The Championships. Whilst the roof is being opened or closed, play is suspended. The first time the roof was closed during a Wimbledon Championship match was on Monday 29 June 2009, involving Amélie Mauresmo and Dinara Safina. The first full match played and completed under the roof featured Andy Murray and Stan Wawrinka, played on the same date.
The court has a capacity of 15,000. At its south end is the Royal Box, from which members of the Royal Family and other dignitaries watch matches. Centre Court usually hosts the finals and semifinals of the main events, as well as many matches in the earlier rounds involving top-seeded players or local favourites.
The second most important court is No. 1 Court. The court was constructed in 1997 to replace the old No.1 Court, which was adjacent to Centre Court. The old No.1 Court was demolished because its capacity for spectators was too low. The court was said to have had a unique, more intimate atmosphere and was a favourite of many players. Construction of a new retractable roof on the No.1 Court began after the 2017 Championships and was completed in time for the 2019 championships. The capacity of the stadium also rose by 900 to 12,345.
Since 2009, a new No. 2 Court has been used at Wimbledon with a capacity for 4,000 people. To obtain planning permission, the playing surface is around 3.5m below ground level, ensuring that the single-storey structure is only about 3.5m above ground level, and thus not affecting local views. Plans to build on the current site of Court 13 were dismissed due to the high capacity of games played at the 2012 Olympic Games. The old No.2 Court has been renamed as No.3 Court. The old No.2 Court was known as the "Graveyard of Champions" because many highly seeded players were eliminated there during early rounds over the years, including Ilie Năstase, John McEnroe, Boris Becker, Andre Agassi, Pete Sampras, Martina Hingis, Venus Williams, Serena Williams and Maria Sharapova. The court has a capacity of 2,192 + 770 standing. In 2011 a new No.3 Court and a new Court 4 were unveiled on the sites of the old No.2 and 3 courts.
Because of the summer climate in southern England, Wimbledon employs 'Court Attendants' each year, who work to maintain court conditions. Their principal responsibility is to ensure that the courts are quickly covered when it begins to rain, so that play can resume as quickly as possible once the referees decide to uncover the courts. The outer court attendants are mainly university students working to make summer money. Centre Court is covered by full-time groundstaff, however.
At the northern end of the grounds is a giant television screen on which important matches are broadcast to fans inside the grounds without tickets to the relevant court. Fans watch from a gently inclined area of grass officially known as the Aorangi Terrace. When British players do well at Wimbledon, this area attracts fans for them, and is often renamed after them by the press: Greg Rusedski's followers convened at "Rusedski Ridge", and Tim Henman has had the hill nicknamed Henman Hill. As both of them have now retired and Andy Murray is the most successful current British player, the hill is occasionally referred to as "Murray Mound" or "Murrayfield", as a reference to his Scottish heritage and the Scottish rugby ground of the same name, but this has largely failed to catch on – the area is still usually referred to as Henman Hill. None of these nicknames are the official name.
1913 suffragette terror attack
An attempt was made to destroy the grounds in 1913, during the suffragette bombing and arson campaign. The suffragettes, as part of their campaign for women's votes before the First World War, had begun carrying out politically-motivated arson and bombings across the country. On the night of 27 February 1913, a suffragette woman "between the ages of 30–35" was arrested within the grounds, after being spotted by a groundsman climbing over a hedge at around midnight. She was found to have with her some paraffin and wood shavings, for the purpose of setting fires in the grounds. The woman refused to give her name or any information to the police and was later sentenced to two months' imprisonment.
Bank of England Sports Centre
The qualifying matches, prior to the main draw, take place at the Bank of England Sports Ground, in Roehampton, from the All England Club.
Traditions
Social commentator Ellis Cashmore describes Wimbledon as having "a David Niven-ish propriety", in trying to conform to the standards of behaviour regarded as common in the 1950s. Writer Peter York sees the event as representing a particular white, upper middle class, affluent type of Britishness, describing the area of Wimbledon as "a southern, well off, late-Victorian suburb with a particular social character". Cashmore has criticised the event for being "remote and insulated" from the changing multicultural character of modern Britain, describing it as "nobody's idea of all-things-British".
Ball boys and ball girls
In the championship games, ball boys and girls, known as BBGs, play a crucial role in the smooth running of the tournament, with a brief that a good BBG "should not be seen. They should blend into the background and get on with their jobs quietly."
From 1947 ball boys were recruited from Goldings, the only Barnardos school to provide them. Prior to this, from the 1920s onwards, the ball boys came from The Shaftesbury Children's Home.
Since 1969, BBGs have been drawn from local schools. Traditionally, Wandsworth Boys School in Sutherland Grove, Southfields and Mayfield Girls School on West Hill in Wandsworth, both now defunct, were the schools of choice for selection of BBGs. This was possibly owing to their proximity to the club. Since 2008 they have been drawn from schools in the London boroughs of Merton, Sutton, Kingston, and Wandsworth, as well as from Surrey. BBGs have an average age of 15, being drawn from the school years nine and ten. They serve for one, or if re-selected, for up to five tournaments, up to year thirteen.
Starting in 2005, BBGs work in teams of six, two at the net, four at the corners, and teams rotate one hour on court, one hour off, (two hours depending on the court) for the day's play. Teams are not told which court they will be working on the day, to ensure the same standards across all courts. With the expansion of the number of courts, and lengthening the tennis day, as of 2008, the number of BBGs required is around 250. Starting on the second Wednesday, the number of BBGs is reduced due to the decrease in the number of matches per day, leaving around 80 on the final Sunday. Each BBG receives a certificate, a can of used balls, a group photograph and a programme when leaving. BBG service is paid, with a total of £160-£250 being paid to each ball boy or girl after the 13-day period, depending on the number of days served, around £17 per day. Every BBG keeps their kit. BBG places are split 50:50 between boys and girls, with girls having been included since 1977, appearing on centre court since 1985.
Prospective BBGs are first nominated by their school headteacher, to be considered for selection. To be selected, a candidate must pass written tests on the rules of tennis, and pass fitness, mobility and other suitability tests, against initial preliminary instruction material. Successful candidates then commence a training phase, starting in February, in which the final BBGs are chosen through continual assessment. As of 2008, this training intake was 600. The training includes weekly sessions of physical, procedural and theoretical instruction, to ensure that the BBGs are fast, alert, self-confident and adaptable to situations. As of 2011, early training occurs at the Wimbledon All England Lawn Tennis Club Covered Courts, to the side of the Grounds, and then moves to outside courts (8, 9, 10) the week before the Championships to ensure that BBGs gain a feel of the grass court.
Colours and uniforms
Dark green and purple are the traditional Wimbledon colours. However, all tennis players participating in the tournament are required to wear all-white or at least almost all-white clothing, a long-time tradition at Wimbledon. Wearing white clothing with some colour accents is also acceptable, provided the colour scheme is not that of an identifiable commercial brand logo (the outfitter's brand logo being the sole exception). Controversy followed Martina Navratilova's wearing branding for "Kim" cigarettes in 1982. Green clothing was worn by the chair umpire, linesmen, ball boys and ball girls until the 2005 Championships; however, beginning with the 2006 Championships, officials, ball boys and ball girls were dressed in new navy blue- and cream-coloured uniforms from American designer Ralph Lauren.
Referring to players
By tradition, the "Men's" and "Women's" competitions are referred to as "Gentlemen's" and "Ladies'" competitions at Wimbledon. The junior competitions are referred to as the "Boys'" and "Girls'" competitions.
Prior to 2009, female players were referred to by the title "Miss" or "Mrs" on scoreboards. On Wimbledon's Champions Board, married female players were referred to by their husband's name up until 2019, as can be seen in the "Ladies Names" list. For the first time during the 2009 tournament, players were referred to on scoreboards by both their first and last names.
The title "Mr" is not used for male players who are professionals on scoreboards but is retained for amateurs, although chair umpires refer to players as "Mr" when they use the replay challenge. The chair umpire will say "Mr <surname> is challenging the call..." and "Mr. <surname> has X challenges remaining." Up until 2018, the chair umpire said "Miss"/"Mrs" <surname> when announcing the score of the Ladies' matches. However, the chair umpire no longer calls "Miss"/"Mrs" <surname> when announcing the score, since 2019.
If a match is being played with two competitors of the same surname (e.g. Venus and Serena Williams, Bob and Mike Bryan), the chair umpire will specify to whom they are referring by stating the player's first name and surname during announcements (e.g. "Game, Venus Williams", "Advantage, Mike Bryan").
Royal family
Previously, players bowed or curtsied to members of the royal family seated in the Royal Box upon entering or leaving Centre Court. However, in 2003, All England Club president Prince Edward, Duke of Kent decided to discontinue the tradition. Now, players are required to bow or curtsy only if the Prince of Wales or the Queen is present, as was in practice during the 2010 Championships when the Queen was in attendance at Wimbledon on 24 June.
On 27 June 2012, Roger Federer said in his post-match interview that he and his opponent had been asked to bow towards the Royal Box as Prince Charles and his wife were present, saying that it was not a problem for him.
Services stewards
Prior to the Second World War, members of the Brigade of Guards and retired members of the Royal Artillery performed the role of stewards. In 1946 the AELTC offered employment to wartime servicemen returning to civilian life during their demobilisation leave. Initially, this scheme extended only to the Royal Navy, followed by the British Army in 1947 and the Royal Air Force in 1949. In 1965 London Fire Brigade members joined the ranks of stewards. The service stewards, wearing uniform, are present in Centre Court and No.'s 1, 2, 3, 12 and 18 courts. In 2015, 595 Service and London Fire Brigade stewards attended. Only enlisted members of the Armed Forces may apply for the role, which must be taken as leave, and half of each year's recruits must have stewarded at Wimbledon before. The AELTC pays a subsistence allowance to servicemen and women working as stewards to defray their accommodation costs for the period of the Championships. The Service Stewards are not to be confused with the 185 Honorary Stewards.
Tickets
The majority of centre and show court tickets sold to the general public have since 1924 been made available by a public ballot that the All England Lawn Tennis and Croquet Club holds at the start of the year. The ballot has always been substantially oversubscribed. Successful applicants are selected at random by a computer. The most recent figures from 2011 suggested there were four applicants to every ballot ticket. Applications must be posted to arrive at the AELTC by the last day of December in the year prior to the tournament. Seats and days are allocated randomly and ballot tickets are not transferable.
The All England Club, through its subsidiary The All England Lawn Tennis Ground plc, issues debentures to tennis fans every five years to raise funds for capital expenditure. Fans who invest thus in the club receive a pair of tickets for every day of the Wimbledon Championships for the five years the investment lasts. Only debenture holders are permitted to sell on their tickets to third parties and demand for debentures has increased in recent years, to such an extent that they are even traded on the London Stock Exchange.
Wimbledon and the French Open are the only Grand Slam tournaments where fans without tickets for play can queue up and still get seats on the three show courts on the day of the match. Sequentially numbered queue cards were introduced in 2003. From 2008, there is a single queue, allotted about 500 seats for each court. When they join the queue, fans are handed queue cards. Anyone who then wishes to leave the queue temporarily, even if in possession of a queue card, must agree their position with the others nearby in the queue and/or a steward.
To get access to the show courts, fans normally have to queue overnight. This is done by fans from all over the world and, although considered vagrancy, is part of the Wimbledon experience in itself. The All-England Club allows overnight queuing and provides toilet and water facilities for campers. Early in the morning when the line moves towards the Grounds, stewards walk along the line and hand out wristbands that are colour-coded to the specific court. The wrist band (and payment) is exchanged at the ticket office for the ticket when the grounds open. General admission to the grounds gives access to the outer courts and is possible without queuing overnight. Tickets returned by people leaving early go on sale at 2:30pm and the money goes to charity. Queuing for the show courts ends after the quarter finals have been completed.
At 2.40pm on Day Seven (Monday 28 June) of the 2010 Championships, the one-millionth numbered Wimbledon queue card was handed out to Rose Stanley from South Africa.
Sponsorship
Wimbledon is notable for the longest running sponsorship in sports history due to its association with Slazenger who have supplied all tennis balls for the tournament since 1902. Wimbledon has had a sponsorship deal with the Robinsons fruit drink brand since 1935.
Strawberries and cream
Strawberries and cream are traditionally eaten by spectators at Wimbledon and have become culturally synonymous with the tournament. In 2017, fans consumed 34,000kg (33 imperial tons) of British strawberries and 10,000 litres (2,200 imperial gallons) of cream.
Media
Radio Wimbledon
Until 2011 when its contract ended, Radio Wimbledon could be heard within a five-mile radius on 87.7 FM, and also online. It operated under a Restricted Service Licence. Presenters included Sam Lloyd and Ali Barton. Typically they worked alternate four-hour shifts until the end of the last match of the day. Reporters and commentators included Gigi Salmon, Nick Lestor, Rupert Bell, Nigel Bidmead, Guy Swindells, Lucie Ahl, Nadine Towell and Helen Whitaker. Often they reported from the "Crow's Nest", an elevated building housing the Court 3 and 4 scoreboards which affords views of most of the outside courts. Regular guests included Sue Mappin. In later years Radio Wimbledon acquired a second low-power FM frequency (within the grounds only) of 96.3 FM for uninterrupted Centre Court commentary, and, from 2006, a third for coverage from No. 1 Court on 97.8 FM. Hourly news bulletins and travel (using RDS) were also broadcast.
Television coverage
Beginning with the 2018 tournament, an in-house operation known as Wimbledon Broadcasting Services (WBS) has served as the official host broadcaster of the tournament, replacing BBC Sport.
United Kingdom
Since 1937 the BBC has broadcast the tournament on television in the United Kingdom. Between 1956 and 1968 The Championships were also covered by the ITV Network, but since 1969 the BBC has had a monopoly. The matches covered are primarily split between its two main terrestrial channels, BBC One and BBC Two, and their Red Button service. This can result in live matches being moved across all 3 channels. The BBC holds the broadcast rights for Wimbledon until 2027. During the days of British Satellite Broadcasting, its sports channel carried extra coverage of Wimbledon for subscribers. One of the most notable British commentators was Dan Maskell, who was known as the BBC's "voice of tennis" until his retirement in 1991. John Barrett succeeded him in that role until he retired in 2006. Current commentators working for the BBC at Wimbledon include British ex-players Andrew Castle, John Lloyd, Tim Henman, Greg Rusedski, Samantha Smith and Mark Petchey; tennis legends such as John McEnroe, Tracy Austin, Boris Becker and Lindsay Davenport; and general sports commentators including David Mercer, Barry Davies, Andrew Cotter and Nick Mullins. The coverage is presented by Sue Barker (live) and Claire Balding (highlights). Previous BBC presenters include Des Lynam, David Vine, John Inverdale and Harry Carpenter.
The Wimbledon Finals are obliged to be shown live and in full on terrestrial television (BBC Television Service, ITV, Channel 4, or Channel 5) by government mandate. Highlights of the rest of the tournament must be provided by terrestrial stations; live coverage (excepting the finals) may be sought by satellite or cable TV.
The BBC was forced to apologise after many viewers complained about "over-talking" by its commentary team during the TV coverage of the event in 2011. It said in a statement that views on commentary were subjective but that they "do appreciate that over-talking can irritate our audience". The BBC added that it hoped it had achieved "the right balance" across its coverage and was "of course sorry if on occasion you have not been satisfied". Tim Henman and John McEnroe were among the ex-players commentating.
Wimbledon was also involved in a piece of television history, when on 1 July 1967 the first official colour television broadcast took place in the UK. Four hours live coverage of the 1967 Championships was shown on BBC Two, which was the first television channel in Europe to regularly broadcast in colour. Footage of that historic match no longer survives, however, the Gentlemen's Final of that year is still held in the BBC archives because it was the first Gentlemen's Final transmitted in colour. The tennis balls used were traditionally white, but were switched to yellow in 1986 to make them stand out for colour television. Since 2007, Wimbledon matches have been transmitted in high-definition, originally on the BBC's free-to-air channel BBC HD, with continual live coverage during the tournament of Centre Court and Court No. 1 as well as an evening highlights show Today at Wimbledon. Coverage is now shown on BBC One and Two's HD feeds. Beginning 2018, all centre court matches are televised in 4K ultra-high-definition.
The BBC's opening theme music for Wimbledon was composed by Keith Mansfield and is titled "Light and Tuneful". A piece titled "A Sporting Occasion" is the traditional closing theme, though nowadays coverage typically ends either with a montage set to a popular song or with no music at all. Mansfield also composed the piece "World Champion", used by NBC during intervals (change-overs, set breaks, etc.) and at the close of broadcasts throughout the tournament.
Ireland
In Ireland, RTÉ broadcast the tournament during the 1980s and 1990s on their second channel RTÉ Two, they also provided highlights of the games in the evening. The commentary provided was given by Matt Doyle a former Irish-American professional tennis player and Jim Sherwin a former RTÉ newsreader. Caroline Murphy was the presenter of the programme. RTÉ made the decision in 1998 to discontinue broadcasting the tournament due to falling viewing figures and the large number of viewers watching on the BBC. From 2005 until 2014 TG4 Ireland's Irish-language broadcaster provided coverage of the tournament. Live coverage was provided in the Irish language while they broadcast highlights in English at night.
In 2015 Wimbledon moved to pay TV broadcaster Setanta Sports under a 3-year agreement. Its successor, Eir Sport, currently hosts broadcasting rights in Ireland.
Americas
In the United States, ABC began showing taped highlights of the Wimbledon Gentlemen's Singles Final in the 1960s on its Wide World of Sports series. NBC began a 43-year run of covering Wimbledon in 1969, with same-day taped (and often edited) coverage of the Gentlemen's Singles Final. In 1979, the network began carrying the Gentlemen's and Ladies' Singles Finals live. For the next few decades, Americans made a tradition of NBC's "Breakfast at Wimbledon" specials at weekends. Live coverage started early in the morning (the US being a minimum of 5 hours behind the UK) and continued well into the afternoon, interspersed with commentary and interviews from Bud Collins, whose tennis acumen and famous patterned trousers were well known to tennis fans in the US. Collins was sacked by NBC in 2007, but was promptly hired by ESPN, the cable home for The Championships in the States. For many years NBC's primary Wimbledon host was veteran broadcaster Dick Enberg.
From 1975 to 1999, premium channel HBO carried weekday coverage of Wimbledon. Hosts included Jim Lampley, Billie Jean King, Martina Navratilova, John Lloyd and Barry MacKay among others. ESPN took over as the cable-television partner in 2003.
The AELTC grew frustrated with NBC's policy of waiting to begin its quarterfinal and semifinal coverage until after the conclusion of Today at 10 a.m. local, as well as broadcasting live only to the Eastern Time Zone and using tape-delay in all others. NBC also held over high-profile matches for delayed broadcast in its window, regardless of any ongoing matches. In one notorious incident in 2009, ESPN2's coverage of the Tommy Haas-Novak Djokovic quarterfinal was forced off the air nationwide when it ran past 10 a.m. Eastern, after which NBC showed the conclusion of the match on tape only after presenting the previous Ivo Karlović-Roger Federer quarterfinal in full. Beginning with the 2012 tournament, coverage moved to ESPN and ESPN2, marking the second major tennis championship (after the Australian Open) where live coverage is exclusively on pay television, while ESPN Deportes provide the coverage in Spanish. The finals are also broadcast tape-delayed on ABC. On 9 July 2021, ESPN and AELTC reached an agreement to extend the coverage for 12 years, beginning from 2024 until 2035. This agreement is including live coverage on ABC of play on the middle weekend which begins in 2022, after AELTC announces will no longer schedule a rest day on its middle Sunday.
Taped coverage using the world feed is aired in primetime and overnights on Tennis Channel and is branded Wimbledon Primetime.
In Canada, coverage of Wimbledon is exclusively carried by TSN and RDS, which are co-owned by Bell Media and ESPN. Prior to 2012, CBC Television and SRC were the primary broadcaster of Wimbledon for Canada, and its live coverage of the tournament predated "Breakfast at Wimbledon" by over a decade, Canada being at least four hours from its fellow Commonwealth realm.
In Mexico, the Televisa family of networks has aired Wimbledon since the early 1960s. Presently, most weekend matches are broadcast through Canal 5 with the weekday matches broadcast on the Televisa Deportes Network. As Mexico is six hours behind the U.K., some Canal 5 affiliates air the weekend matches as the first program of the day after sign-on. Although Mexico had begun broadcasting in colour in 1962, Wimbledon continued to air in black and white in Mexico until colour television came to the United Kingdom in 1967.
In most of the remainder of Latin America, Wimbledon airs on ESPN, as do the other Grand Slam tournaments. In Brazil, SporTV has exclusive rights to the broadcast.
Other countries
In several European countries, Wimbledon is shown live on Eurosport 1, Eurosport 2 and the Eurosport Player. Although there are some exceptions, as in Denmark, where the Danish TV2 holds the right to show matches until 2022. In the Netherlands Center Court is shown live on Eurosport 1 and all other courts are shown live on the Eurosport Player. But Court One is covered live on Ziggo Sport/Ziggo Sport Select.
In Australia, the free-to-air Nine Network covered Wimbledon for almost 40 years but decided to drop their broadcast following the 2010 tournament, citing declining ratings and desire to use money saved to bid on other sports coverage. In April 2011, it was announced that the Seven Network, the then-host broadcaster of the Australian Open, along with its sister channel 7Two would broadcast the event from 2011. Pay television network Fox Sports Australia also covers the event. Free-to-air coverage returned to Nine Network in 2021. In India and its Subcontinental region, it is broadcast on Star Sports. In Pakistan it is broadcast on PTV Sports.
Coverage is free-to-air in New Zealand through TVNZ One, beginning each night at 11 pm (midday in London). In 2017 their new channel, TVNZ Duke (also free-to-air), carried an alternative to the main feed, including (for example) matches on outside courts involving New Zealand players.
Fox Sports Asia holds broadcasting rights across Southeast Asia.
Most matches are also available for viewing through internet betting websites and other live streaming services, as television cameras are set up to provide continuous coverage on nearly all the courts.
Trophies and prize money
Trophies
The Gentlemen's Singles champion is presented with a silver gilt cup 18.5inches (about 47cm) in height and 7.5inches (about 19cm) in diameter. The trophy has been awarded since 1887 and bears the inscription: "All England Lawn Tennis Club Single Handed Championship of the World". The actual trophy remains the property of the All England Club in their museum, so the champion receives a three-quarter size replica of the Cup bearing the names of all past Champions (height 13.5 inches, 34cm).
The Ladies' Singles champion is presented with a sterling silver salver commonly known as the "Venus Rosewater Dish", or simply the "Rosewater Dish". The salver, which is 18.75inches (about 48cm) in diameter, is decorated with figures from mythology. The actual dish remains the property of the All England Club in their museum, so the champion receives a miniature replica bearing the names of all past Champions. From 1949 to 2006 the replica was 8 inches in diameter, and since 2007 it has been a three-quarter size replica with a diameter of 13.5 inches.
The winner of the Gentlemen's Doubles, Ladies' Doubles, and Mixed Doubles events receive silver cups. A trophy is awarded to each player in the Doubles pair, unlike the other Grand Slam tournaments where the winning Doubles duo shares a single trophy. The Gentlemen's Doubles silver challenge cup was originally from the Oxford University Lawn Tennis Club and donated to the All England Club in 1884. The Ladies' Doubles Trophy, a silver cup and cover known as The Duchess of Kent Challenge Cup, was presented to the All England Club in 1949 by The Duchess of Kent. The Mixed Doubles Trophy is a silver challenge cup and cover presented to the All England Club by the family of two-time Wimbledon doubles winner S.H. Smith.
The runner-up in each event receives an inscribed silver plate. The trophies are usually presented by the President of the All England Club, The Duke of Kent.
Prize money
Prize money was first awarded in 1968, the year that professional players were allowed to compete in the Championships for the first time. Total prize money was £26,150; the winner of the men's title earned £2,000 while the women's singles champion received £750. In money as of 2018, £2,000 as of 1968 equals £34,600 however. In 2007, Wimbledon and French Open became the last grand slam tournaments to award unequal prize money to women and men.
In 2009, a total of £12,500,000 in prize money was awarded with the singles champions receiving £850,000 each, an increase of 13.3 percent on 2008. In 2010 total prize money increased to £13,725,000, and the singles champions received £1,000,000 each. A further increase of 6.4% in 2011 resulted in a total prize money amount of £14,600,000. Both male and female singles champions prize money increased to £1,100,000, a rise of 10%. The 2012 Championships offered total prize money of £16,060,000, an increase of 10.0% from 2011. The bulk of the increases were given to players losing in earlier rounds. This move was in response to the growing angst among lower-ranked players concerning the inadequacy of their pay. Sergiy Stakhovsky, a member of the ATP Player Council and who was at the time ranked 68th, was among the most vocal in the push for higher pay for players who bow out in the earlier rounds. In an interview Stakhovsky intimated that it is not uncommon for lower-ranked players to be in the negative, for certain tour events, if their results were not stellar. This issue gained the attention of the men's "Big Four"—Novak Djokovic, Roger Federer, Andy Murray, and Rafael Nadal—as well as the Championships.
In 2013 total prize money was further increased by about 40% from 2012 to £22,560,000. The losers in the earlier singles rounds of the tournament saw a highest 62% increase in their pay while the total prize money of the doubles increased by 22%. The prize money for participants of the qualifying matches saw an increase of 41%. Sergiy Stakhovsky, a member of the ATP Player Council, was the loudest voice for this increase. The 2015 prize money was £1,880,000 each for the Gentlemen's and Ladies' Singles winners, £340,000 each pair for the Gentlemen's and Ladies' Doubles winners, and £100,000 per pair for the Mixed Doubles winners. The total prize money awarded is £26,750,000 up 7% from the £25,000,000 in 2014. The 2016 Wimbledon Championships saw prize money for the Gentlemen's and Ladies' Singles winners reach £2,000,000 for the first time. The winning pair of the Gentlemen's and Ladies' Doubles received £350,000, a £10,000 increase from 2015. £100,000 was awarded to the winning pair of the Mixed Doubles competition.
In 2016, the total prize money of £28,100,000 was a 5% increase on the 2015 prize money.
In 2017, the total prize money rose by 12.5% to £31,600,000, with the Gentlemen's and Ladies' Singles Champions each receiving £2.2 million, a 10% increase from £2.0 million in 2016. In 2019 prize money was increased to £38,000,000.
Doubles prize money is per team.
Ranking points
Ranking points for the ATP and WTA have varied at Wimbledon through the years but at present singles players receive the following points:
Champions
Past champions
Gentlemen's singles
Ladies' singles
Gentlemen's doubles
Ladies' doubles
Mixed doubles
Open Era champions
Current champions
Most recent finals
Records
Gentlemen since 1877
Ladies since 1884
Miscellaneous
See also
2012 Summer Olympics venues
List of British finalists at Grand Slam tennis tournaments
Wimbledon (film)
Wimbledon Effect
Lists of champions
List of Wimbledon Open Era champions (Open Era, all events)
List of Wimbledon gentlemen's singles champions
List of Wimbledon ladies' singles champions
List of Wimbledon gentlemen's doubles champions
List of Wimbledon ladies' doubles champions
List of Wimbledon mixed doubles champions
List of Wimbledon singles finalists during the open era, records and statistics
Other Grand Slam tournaments
Australian Open
French Open
US Open
Notes
References
Further reading
External links
1877 establishments in England
Annual events in London
Annual sporting events in the United Kingdom
Wimbledon
Grass court tennis tournaments
July sporting events
Major tennis tournaments
Recurring sporting events established in 1877
Sport in the London Borough of Merton
Tennis in London
Tennis tournaments in England
Wimbledon, London |
null | null | Friedrich Wilhelm von Steuben | eng_Latn | Friedrich Wilhelm August Heinrich Ferdinand von Steuben (born Friedrich Wilhelm Ludolf Gerhard Augustin Louis von Steuben; September 17, 1730 – November 28, 1794), also referred to as Baron von Steuben (), was a Prussian military officer who played a leading role in the American Revolutionary War by reforming the Continental Army into a disciplined and professional fighting force. His contributions marked a significant improvement in the performance of American troops, and he is subsequently regarded as one of the fathers of the United States Army.
Born into a military family, Steuben was exposed to war from an early age; at 14 years old, he observed his father directing Prussian engineers in the 1744 siege of Prague. He enlisted in the Prussian Army at age 16 or 17, which was considered the most professional and disciplined in Europe. During his 17 years of miliary service, Steuben took part in several battles in the Seven Years' War (1756–63), rose to the rank of captain, and became aide-de-camp to Prussian King Frederick the Great, who was renowned for his military prowess and strategy. Steuben's career culminated in his attendance of Frederick's elite school for young military officers, after which he was abruptly discharged from the army in 1763, allegedly by the machinations of a rival.
Steuben spent 11 years as court chamberlain to the prince of Hohenzollern-Hechingen, a small German principality. In 1769, the Duchess of Wurttemburg, a niece of Frederick, named him to the chivalric Order of Fidelity, a meritorious award that conferred the title, Freiherr, or "free lord"; in 1771, his service to Hollenzollern-Hechingen earned him the title baron. In 1775, as the American Revolution had begun, Steuben saw a reduction in his salary and sought some form of military work; unable to find employment in peacetime Europe, he joined the American war effort through mutual French contacts with American diplomats, most notably ambassador to France Benjamin Franklin. Due to his military exploits, and his willingness to serve the Americans without compensation, Steuben made a positive impression on both Congress and General George Washington, who appointed him as temporary Inspector General of the Continental Army.
Appalled by the state of American forces, Steuben took the lead in teaching soldiers the essentials of military drills, tactics, and discipline based on Prussian techniques. He wrote Regulations for the Order and Discipline of the Troops of the United States, which remained the army's drill manual for decades, and continues to influence modern U.S. army manuals. Steuben also addressed widespread administrative waste and graft, helping save desperately needed supplies and funds. Impressed by these results, which began bearing fruit on the battlefield, in 1778, Congress, on Washington's recommendation, commissioned Steuben as Inspector General with the rank of Major General. He served the remainder of the war as Washington's chief of staff and one of his most trusted advisors.
After the war, Steuben was made a U.S. citizen and granted a large estate in New York in reward for his service. In 1780, he was elected a member of the American Philosophical Society, a learned society that included most of the nation's most prominent founding fathers.
Early life and education
Baron von Steuben was born in the fortress town of Magdeburg, Germany, on September 17, 1730, the son of Royal Prussian Engineer Capt. Wilhelm von Steuben and his wife, Elizabeth von Jagvodin. When his father entered the service of Empress Anna of Russia, young Friedrich went with him to Crimea and then to Kronstadt, staying until the Russian war against the Turks under General Burkhard Christoph von Münnich. In 1740, Steuben's father returned to Prussia and Friedrich was educated in the garrison towns Neisse and Breslau by Jesuits. Despite his education by a Catholic order, von Steuben remained critical of Roman Catholicism. Originally, von Steuben's family were Protestants in the Kingdom of Prussia, and after his emigration to America he became a member of the Reformed German Church, a Reformed congregation in New York. It is said that at age 14 he served as a volunteer with his father in one of the campaigns of the War of the Austrian Succession.
First military service
Baron von Steuben joined the Prussian Army at age 17. He served as a second lieutenant during the Seven Years' War in 1756, and was wounded at the 1757 Battle of Prague. He served as adjutant to the free battalion of General Johann von Mayr and was promoted to first lieutenant in 1759. In August 1759 he was wounded a second time at the Battle of Kunersdorf. In the same year, he was appointed deputy quartermaster at the general headquarters. In 1761 he became adjutant of the Major General Von Knobloch upon being taken prisoner by the Russians at Treptow. He subsequently attained the rank of captain and served as aide-de-camp to Frederick the Great; in 1762 he was one of 13 young officers chosen to participate in a special course of instruction delivered by the king himself.
Upon the reduction of the army at the end of the war, in 1763, Steuben was one of many officers who found themselves unemployed. Towards the end of his life, Steuben indicated in a letter that "an inconsiderate step and an implacable personal enemy" led to his leaving the Prussian army.
Service in Hohenzollern-Hechingen
In 1764, Steuben became Hofmarschall to Fürst Josef Friedrich Wilhelm of Hohenzollern-Hechingen, a post he held until 1777. In 1769 the Duchess of Württemberg, niece of Frederick the Great, presented him with the Cross of the Order of De la Fidelite. In 1771 he began to use the title baron. That same year he accompanied the prince to France, hoping to borrow money. Failing to find funds, they returned to Germany in 1775, deeply in debt.
In 1763, Steuben had been formally introduced to the future French Minister of War, Claude Louis, Comte de Saint-Germain, in Hamburg. They met again in Paris in 1777. The Count, fully realizing the potential of an officer with Prussian general staff training, introduced him to Benjamin Franklin. Franklin, however, was unable to offer Steuben a rank or pay in the American army. The Continental Congress had grown tired of foreign mercenaries coming to America and demanding a high rank and pay. Promoting these men over qualified American officers caused discontent in the ranks. Steuben would have to go to America strictly as a volunteer and present himself to Congress. Steuben left these first meetings in disgust and returned to Prussia. After unknown events, Steuben discharged from his position as a captain and traveled to Paris. It is speculated that he was, or was accused to be, a homosexual. It is unknown whether or not this occurred, and either way, no charges were pursued.
Upon the Count's recommendation, Steuben was introduced to future president George Washington by means of a letter from Franklin as a "Lieutenant General in the King of Prussia's service", an exaggeration of his actual credentials that appears to be based on a mistranslation of his service record. He was advanced travel funds and left Europe from Marseilles on Friday, September 26, 1777, on board the frigate Flamand.
American Revolution
The Baron, his Italian Greyhound Azor (which he took with him everywhere), his young aide-de-camp Louis de Pontière, his military secretary, Peter Stephen Du Ponceau (then called Pierre Etienne Du Ponceau), and two other companions, reached Portsmouth, New Hampshire, on December 1, 1777, where they were almost arrested for being British because Steuben had mistakenly outfitted them in red uniforms. They were extravagantly entertained in Boston. On February 5, 1778, Steuben and his party arrived in York, Pennsylvania, where the Continental Congress had relocated after being ousted from Philadelphia by the British advance. Arrangements were made for Steuben to be paid following the successful completion of the war according to his contributions. He arrived at Valley Forge on February 23, 1778, and reported for duty as a volunteer. One soldier's first impression of the Baron was "of the ancient fabled God of War ... he seemed to me a perfect personification of Mars. The trappings of his horse, the enormous holsters of his pistols, his large size, and his strikingly martial aspect, all seemed to favor the idea. He turned the volunteers into a great army."
Inspector General
Washington appointed von Steuben as temporary inspector general. He went out into the camp to talk with the officers and men, inspect their huts, and scrutinize their equipment. Steuben established standards of sanitation and camp layouts that would still be standard a century and a half later. There had previously been no set arrangement of tents and huts. Men relieved themselves where they wished, and when an animal died it was stripped of its meat and the rest was left to rot where it lay. Steuben laid out a plan to have rows for command, officers, and enlisted men. Kitchens and latrines were on opposite sides of the camp, with latrines on the downhill side. There was the familiar arrangement of company and regimental streets.
On May 5, 1778, on General Washington's recommendation, Congress appointed Steuben inspector general of the army, with the rank and pay of major general. The internal administration had been neglected, and no books had been kept either as to supplies, clothing, or men. Steuben became aware of the "administrative incompetence, graft, war profiteering" that existed. He enforced the keeping of exact records and strict inspections. His inspections saved the army an estimated loss of five to eight thousand muskets.
Training program
Steuben picked 120 men from various regiments to form an honor guard for General Washington, and used them to demonstrate military training to the rest of the troops. These men in turn trained other personnel at regimental and brigade levels. Steuben's eccentric personality greatly enhanced his mystique. In full military dress uniform, he twice a day trained the soldiers who, at this point, were themselves greatly lacking in proper clothing.
As he could only speak and write a small amount of English, Steuben originally wrote the drills in German, the military language of Europe at the time. His secretary, Du Ponceau, then translated the drills from German into French, and a secretary for Washington translated it to English. They did this every single night so Washington could command his soldiers in the morning. Colonel Alexander Hamilton and General Nathanael Greene were of great help in assisting Steuben in drafting a training program for the Army. The Baron's willingness and ability to work with the men, as well as his use of profanity (in several different languages), made him popular among the soldiers. It is here he met his close friend and future adopted heir, Captain Benjamin Walker. Within weeks, Walker was Steuben's aide-de-camp.
Steuben introduced a system of progressive training, beginning with the school of the soldier, with and without arms, and going through the school of the regiment. This corrected the previous policy of simply assigning personnel to regiments. Each company commander was made responsible for the training of new men, but actual instruction was done by sergeants specifically selected for being the best obtainable.
In the earlier part of the war, the Americans used the bayonet mostly as a cooking skewer or tool rather than as a fighting instrument. Steuben's introduction of effective bayonet charges became crucial. In the Battle of Stony Point, Continental Army soldiers attacked with unloaded muskets and won the battle solely on Steuben's bayonet training.
The first results of Steuben's training were in evidence at the Battle of Barren Hill, May 20, 1778, and then again at the Battle of Monmouth in June 1778. Steuben, by then serving in Washington's headquarters, was the first to determine that the enemy was heading for Monmouth.
During the winter of 1778–1779, Steuben prepared Regulations for the Order and Discipline of the Troops of the United States, commonly known as the "Blue Book". Its basis was the training plan he had devised at Valley Forge. It was used by the United States Army until 1814, and affected U.S. drills and tactics until the Mexican–American War of 1846.
On May 2, 1779, during the second Middlebrook encampment, a review of the army was held to honor the French minister Conrad Alexandre Gérard de Rayneval and the Spanish diplomat Juan de Miralles. Led by General William Smallwood, four battalions performed precise military formations to demonstrate their mastery of Steuben's training. After the review, about sixty generals and colonels attended a dinner hosted by Steuben in a large tent near his headquarters at the Abraham Staats House.
Southern campaign
In 1780 Steuben sat on the court-martial of the British Army officer Major John André, captured and charged with espionage in conjunction with the defection of General Benedict Arnold. He later traveled with Nathanael Greene, the new commander of the Southern campaign. He quartered in Virginia, since U.S. supplies and soldiers would be provided to the army from there. Steuben would help in the defense of Virginia with approximately 1,000 militia fighting a delaying action in the Battle of Blandford. During the spring of 1781, he aided Greene in the campaign in the South, culminating in the delivery of 450 Virginia Continentals to Lafayette in June.
He was forced to take sick leave, rejoining the army for the final campaign at Yorktown, where his role was as commander of one of the three divisions of Washington's troops. In 1783, General Von Steuben joined General Knox at Vail's Gate, near West Point, in the fall of 1782 and in early 1783 moved to the Verplanck homestead, at Mount Gulian, across the Hudson River from Washington's headquarters in Newburgh. Steuben gave assistance to Washington in demobilizing the army in 1783 as well as aiding in the defense plan of the new nation. In May 1783, Steuben presided over the founding of the Society of the Cincinnati. He was discharged from the military with honor on March 24, 1784.
Final years
Steuben became a U.S. citizen by act of the Pennsylvania legislature in March 1784 and later by the New York authorities in July 1786. With the war over, Steuben resigned from service and first settled with his longtime companion, William North, for whom he created a special room at his retreat he called the Louvre on Manhattan Island, where he became a prominent figure and elder in the German Reformed Church. From 1785 until his death in 1794, he served as president of the German Society of the City of New York, a charitable society founded in 1784 to assist German immigrants.
In 1786 during Shays' Rebellion, under the written name "Belisarius", Steuben criticized the Massachusetts government for being an oligarchy.
On December 23, 1783, the state of New Jersey presented him with the use of an estate in Bergen County now known as Steuben House, which had been confiscated from Loyalist Jan Zabriskie in 1781. Located in the formerly strategic New Bridge Landing, the estate included a gristmill and about of land. Legislators initially conditioned the grant, requiring Steuben to "hold, occupy and enjoy the said estate in person, and not by tenant." Gen. Philemon Dickinson of the New Jersey Militia informed the baron of this gift and responded to his inquiries that "there are on the premises an exceeding good House, an excellent barn, together with many useful outbuildings, all of which I am told, want some repairs...there is...a Grist-mill; a good Orchard, some meadow Ground, & plenty of Wood. The distance from N York by land 15 miles, but you may keep a boat & go from your own door to N York by water—Oysters, Fish & wild fowl in abundance—Possession will be given to you in the Spring, when you will take a view of the premises." Von Steuben spent considerable sums to repair wartime damages to the house and restore its commercial operations under former aide Walker.
On September 5, 1788, the New Jersey Legislature gave Baron von Steuben full title to the former Zabriskie estate. A month later, recognizing his financial embarrassment, Steuben wrote another former aide-de-camp and companion, William North, recognizing: "The Jersey Estate must and is to be sold. Walker is my administrator, all debts are to be paid out of it." On November 6, 1788, Steuben again wrote North (at his new home in Duanesburg, New York), noting "My Jersey Estate is Advertised but not yet Sold, from this Walker Shall immediately pay to you the money, you so generously lend me and all my debts in New-York will be payed. I support my present poverty with more heroism than I Expected. All Clubs and parties are renounced, I seldom leave the House." Steuben eventually sold the New Jersey property to a son of the previous owner, and it remained in the Zabriskie family until 1909. It is the only remaining eighteenth-century building that von Steuben owned.
Von Steuben was present at the first inauguration of George Washington in New York in 1789.
Von Steuben moved upstate and settled in Oneida County on a small estate in the vicinity of Rome, New York, on land granted to him for his military service and where he had spent summers. He was later appointed a regent for what evolved into the University of the State of New York. In 1790, Congress awarded him a pension of $2,500 a year, which he kept until his death.
Personal life and death
Von Steuben died on November 28, 1794, at his estate in Oneida County, and was buried in a grove at what became the Steuben Memorial State Historic Site. The estate became part of the town of Steuben, New York, which was named for him.
Von Steuben had arrived in the United States with his 17-year-old secretary, Peter Stephen Du Ponceau. At Valley Forge, he began close relationships with Benjamin Walker and William North, then both military officers in their 20s. Von Steuben formally adopted Walker and North and made them his heirs. It is likely that these "extraordinary intense emotional relationships" were romantic, and, given Steuben's reported earlier behavior, it has been suggested it would have been out-of-character for him if they were not. Some 21st-century publications, have embraced him as "a brilliant man", "Godly Man of Valor", or as "the great man who saved the American Revolution".
In any case, von Steuben never married and had no children. He did not care much for his European relatives. Thus, he left his estate to his companions and aides-de-camp, Captain Benjamin Walker and Major General William North, with whom he had had an "extraordinarily intense emotional relationship ... treating them as surrogate sons." A third young man, John W. Mulligan (1774–1862), who also considered himself one of von Steuben's "sons," inherited his vast library, collection of maps and $2,500 in cash. Following von Steuben's death, North divided the property bequeathed to him among his military companions.
Legacy
Generally, Von Steuben Day takes place in September in many cities throughout the United States. It is often considered the German-American event of the year. Participants march, dance, wear German costumes and play German music, and the event is attended by millions of people. The German-American Steuben Parade is held annually in September in New York City. It is one of the largest parades in the city and is traditionally followed by an Oktoberfest in Central Park as well as celebrations in Yorkville, Manhattan, a historically German section of New York City. The German-American Steuben Parade has been taking place since 1958. Chicago also hosts a von Steuben Day parade, which is featured in the U.S. film Ferris Bueller's Day Off. Philadelphia hosts a smaller Steuben Parade in the Northeast section of the city.
The Steuben Society was founded in 1919 as "an educational, fraternal, and patriotic organization of American citizens of German background". In the difficult post-World War I years the Society helped the German-American community to reorganize. It is now one of the largest organizations for Americans of German descent.
A warship, a submarine, and an ocean liner (later pressed into military service) were named in von Steuben's honor. In World War I the captured German ship was renamed as USS Von Steuben, and in World War II there was the Dampfschiff General von Steuben, an ill-fated German luxury passenger ship which was turned into an armed transport ship during the war. During the Cold War, the U.S. Navy submarine was named for him.
Several locations in the United States are also named Steuben, most of them in his honor. Examples include Steuben County, New York, Steuben County, Indiana, and the city of Steubenville, Ohio. Several buildings are named for Steuben, among them Von Steuben Metropolitan High School in Chicago, Illinois, as well as one of the cadet barracks buildings at Valley Forge Military Academy and College.
Von Steuben was one of four European military leaders who assisted the U.S. cause during the Revolution and was honored with a statue in Lafayette Square, just north of the White House, in Washington, D.C. The statue by Albert Jaegers was dedicated in 1910. A copy was dedicated in Potsdam, Germany, in 1911, and destroyed during World War II. A new cast was given in honor of German-American friendship in 1987, and to celebrate the 750th anniversary of the founding of Berlin. It was installed in the Dahlem district, in what had been the U.S. sector of the formerly divided city. An additional cast is in Steuben's home town of Magdeburg. Statues of Steuben by J. Otto Schweizer can be found in Valley Forge, Pennsylvania, and Utica, New York, in addition to an equestrian statue by Schweizer in Milwaukee, Wisconsin. A bust of Steuben is in the garden of the German Embassy in Washington, D.C.
The Steuben House, presented to Steuben as a gift for his services in the Continental Army, is located at New Bridge Landing in River Edge, New Jersey. The house and surrounding farmland were seized in 1781 from a Loyalist family. The house looks much as it did after Steuben renovated it. The State of New Jersey took possession of the historic mansion and one acre of ground for $9,000 on June 27, 1928. It was opened as a public museum in September 1939. The Bergen County Historical Society opens the building for special events. It is under the jurisdiction of the Historic New Bridge Landing Park Commission. It was added to the National Register of Historic Places on December 18, 1970 for its significance in architecture and military history.
Franciscan University of Steubenville also honors their city’s namesake by having The Barons as their moniker and mascot.
Other tributes include Steuben Field, the stadium of the Hamilton College football team. Von Steuben, acting as Alexander Hamilton's surrogate, laid the cornerstone of the school.
The various depictions of Steuben in popular U.S. media include portrayals by Nehemiah Persoff in the 1979 U.S. TV miniseries The Rebels, Kurt Knudson in the 1984 TV miniseries George Washington, being voiced by Austrian-American Arnold Schwarzenegger in the animated series Liberty's Kids, and by David Cross on the "Philadelphia" episode of Drunk History.
In 2007, a popular documentary DVD was released by LionHeart FilmWorks and director Kevin Hershberger titled Von Steuben's Continentals: The First American Army. The 60-minute, live-action documentary details the life, uniforms, camp life, food, weapons, equipment and drill of the Continental soldier 1775–1781, as taught and developed by Baron von Steuben.
See also
Regulations for the Order and Discipline of the Troops of the United States
American Revolutionary War#American strategy
References
Sources
Danckert, Stephen C. "Baron von Steuben and the Training of Armies." Military Review 74 (1994): 29–34 in EBSCO
Steuben Papers, NYHS
The Historic Society of Pennsylvania, Simon Gratz Collection (#250), Case 4, Box 13 William North /Benjamin Walker Letters
Guide to the Friedrich Wilhelm von Steuben Papers at the New-York Historical Society
Further reading
Doyle, Joseph Beatty. Frederick William Von Steuben and the American Revolution: Aide to Washington and Inspector General of the Army (1913).
Palmer, John MacAuley. General von Steuben (Yale UP, 1937).
Whitridge, Arnold. "Baron von Steuben, Washington's Drillmaster." History Today (July 1976) 26#7 pp 429–36.
External links
Friedrich Wilhelm Ludolf Gerhard Augustin Baron von Steuben. Baron Von Steuben's Revolutionary War Drill Manual
Baron Von Steuben: An account of a 19th century visit to America by his German relatives
N.Y. Times: Steuben hosted by the Nation he helped to create
The Society of the Cincinnati
American Revolution Institute
1730 births
1794 deaths
18th-century LGBT people
Barons of Germany
Continental Army generals
Continental Army officers from Germany
Continental Army staff officers
Gay military personnel
German military personnel of the War of the Austrian Succession
Inspectors General of the United States Army
Members of the American Philosophical Society
Prussian emigrants to the United States
Prussian military personnel of the Seven Years' War |
null | null | Waterloo, Texas | eng_Latn | Waterloo was the original name for the city of Austin, Texas, located in Travis County in the central part of the state.
History
After Republic of Texas Vice President Mirabeau B. Lamar visited the area during a buffalo-hunting expedition between 1837 and 1838, he proposed that the republic's capital, then located in Houston, be relocated to an area situated on the north bank of the Colorado River near the present-day Ann W. Richards Congress Avenue Bridge in what is now central Austin. In 1839, the site was officially chosen as the seventh and final location for the capital of the Republic of Texas. It was incorporated under the name "Waterloo". Shortly thereafter, the name was changed to Austin in honor of Stephen F. Austin, the "Father of Texas" and the republic's first secretary of state.
References
External links
Waterloo, Texas from TexasEscapes.com
Waterloo, Texas (Travis County) from https://tshaonline.org
Populated places in Travis County, Texas
History of Austin, Texas
Populated places established in 1839
1839 establishments in the Republic of Texas |
null | null | Cecum | eng_Latn | The cecum or caecum is a pouch within the peritoneum that is considered to be the beginning of the large intestine. It is typically located on the right side of the body (the same side of the body as the appendix, to which it is joined). The word cecum (, plural ceca ) stems from the Latin caecus meaning blind.
It receives chyme from the ileum, and connects to the ascending colon of the large intestine. It is separated from the ileum by the ileocecal valve (ICV) or Bauhin's valve. It is also separated from the colon by the cecocolic junction. While the cecum is usually intraperitoneal, the ascending colon is retroperitoneal.
In herbivores, the cecum stores food material where bacteria are able to break down the cellulose. In humans, the cecum is involved in absorption of salts and electrolytes and lubricates the solid waste that passes into the large intestine.
Structure
Development
The cecum and appendix are formed by the enlargement of the postarterial segment of the midgut loop. The proximal part of the bud grows rapidly to form the cecum. The lateral wall of the cecum grows much more rapidly than the medial wall, with the result that the point of attachment of the appendix comes to lie on the medial side.
History
Etymology
The term cecum comes from Latin (intestinum) caecum, literally 'blind intestine', in the sense 'blind gut' or 'cul de sac'. It is a direct translation from Ancient Greek τυφλὸν (ἔντερον) typhlòn (énteron). Thus the inflammation of the cecum is called typhlitis.
In dissections by the Greek philosophers, the connection between the ileum of the small intestine and the cecum was not fully understood. Most of the studies of the digestive tract were done on animals and the results were compared to human structures.
The junction between the small intestine and the colon, called the ileocecal valve, is so small in some animals that it was not considered to be a connection between the small and large intestines. During a dissection, the colon could be traced from the rectum, to the sigmoid colon, through the descending, transverse, and ascending sections. The cecum is an end point for the colon with a dead-end portion terminating with the appendix.
The connection between the end of the small intestine (ileum) and the start (as viewed from the perspective of food being processed) of the colon (cecum) is now clearly understood, and is called the ileocolic orifice. The connection between the end of the cecum and the beginning of the ascending colon is called the cecocolic orifice.
Clinical significance
A cecal carcinoid tumor is a carcinoid tumor of the cecum. An appendiceal carcinoid tumor (a carcinoid tumor of the appendix) is sometimes found next to a cecal carcinoid.
Neutropenic enterocolitis (typhlitis) is the condition of inflammation of the cecum, primarily caused by bacterial infections.
Over 99% of the bacteria in the gut are anaerobes, but in the cecum, aerobic bacteria reach high densities.
Other animals
A cecum is present in most amniote species, and also in lungfish, but not in any living species of amphibian. In reptiles, it is usually a single median structure, arising from the dorsal side of the large intestine. Birds typically have two paired ceca, as do, unlike other mammals, hyraxes. Parrots do not have ceca.
Most mammalian herbivores have a relatively large cecum, hosting a large number of bacteria, which aid in the enzymatic breakdown of plant matter such as cellulose; in many species, it is considerably wider than the colon. In contrast, obligate carnivores, whose diets contain little or no plant matter, have a reduced cecum, which is often partially or wholly replaced by the appendix. Mammalian species which do not develop a cecum include raccoons, bears, and the red panda.
Many fish have a number of small outpockets, called pyloric ceca, along their intestine; despite the name they are not homologous with the cecum of amniotes, and their purpose is to increase the overall area of the digestive epithelium. Some invertebrates, such as squid, may also have structures with the same name, but these have no relationship with those of vertebrates.
Gallery
See also
McBurney's point
References
External links
Photo at mgccc.cc.ms.us
—"Abdominal organs in situ."
—"The larger intestine."
—"The cecum with the distal portion of the ileum."
—"Incisions of the Cecum"
Video clip of worms in the Cecum
Digestive system
Latin words and phrases |