SEE THIS BLOG & WRITE YOUR OPINION..............FOLLOW MEEEE.........A COMPLETE EDUCATIONAL BLOG.............JOCKS,NEWS,VIDEOS,PICTURES,PROJECT NOTES ARE AVAILABLE HERE.............

India Vision Live News

Monday, October 5, 2009

SHEIK ABOOBACKER BIN AHMAD





























The Muslims of Kerala have been keeping a creative identity in social, political and cultural realms. The social interference fashioned by the Kerala Muslims on the platform of Islamic humanitarianism testifies their vigilant religious life and original social existence. The God-fearing life style of the forgone scholars played a major role in creatively employing the cultural possibilities of the Islamic life. The dynamic moral fiber of the Muslim life in Kerala is imprinted by the vibrant social leadership of the awe-inspiring Ulemas.
The roots of such Organic Intellectual life can be traced in scholars like Sheikh Zainuddin Maqdooms, Omer Qasi, Sayed Alavi of Mampurom, and Ali Musliyar…etc. The public opinion and communal interests of the Muslims of Kerala has been greatly influenced by the brave social interference of these Ulemas, and of course, Kanthapuram A. P. Aboobacker Musliyar is a real inheritor of this Dharma of the rich and scholastic legacy of the Kerala Muslims.
Kanthapuram would be remembered by the future generation on the very fact that he led forward the contemporary Muslim multitude from a chaotic and apprehensive stage to one with confidence and hope. He guided the Kerala Muslim Society by redefining the sense of duty of the Ulama or the learned, and actively involving in the process of social reformation, and thereby played an incomparable role in linking the Islamic society with the mainstream population. He bequeathed great respite on the anxieties of a society, which had been historically sidelined, by presenting the communal issues on the wider perspective of social justice.







CHANDRAYAN-1











Chandrayaan-1, (Sanskrit: चंद्रयान-१, lit: Moon-traveller[3][4] pronunciation (help·info))) was India's first unmanned lunar probe. It was launched by the Indian Space Research Organisation in October 2008, and operated until August 2009. The mission included a lunar orbiter and an impactor. India launched the spacecraft by a modified version of the PSLV, PSLV C11[2][5] on 22 October 2008 from Satish Dhawan Space Centre, Sriharikota, Nellore District, Andhra Pradesh, about 80 km north of Chennai, at 06:22 IST (00:52 UTC).[6] The mission was a major boost to India's space program,[7] as India researched and developed its own technology in order to explore the Moon.[8] The vehicle was successfully inserted into lunar orbit on 8 November 2008.[9]
On 14 November 2008, the Moon Impact Probe separated from the Chandrayaan orbiter at 20:06 and struck the south pole in a controlled manner, making India the fourth country to place its flag on the Moon.[10] The probe impacted near Shackleton Crater at 20:31 ejecting underground soil that could be analysed for the presence of water ice.[11]
The estimated cost for the project was Rs. 386 crore (US$ 80 million).[12]
The remote sensing lunar satellite had a mass of 1,380 kilograms (3,042 lb) at launch and 675 kilograms (1,488 lb) in lunar orbit. [13] It carried high resolution remote sensing equipment for visible, near infrared, and soft and hard X-ray frequencies. Over a two-year period, it was intended to survey the lunar surface to produce a complete map of its chemical characteristics and three-dimensional topography. The polar regions are of special interest as they might contain ice.[14] The lunar mission carries five ISRO payloads and six payloads from other space agencies including NASA, ESA, and the Bulgarian Aerospace Agency, which were carried free of cost.[15]
After suffering from several technical issues including failure of the star sensors and poor thermal shielding, Chandrayaan stopped sending radio signals at 1:30 AM IST on 29 August 2009 shortly after which, the ISRO officially declared the mission over. Chandrayaan operated for 312 days as opposed to the intended two years but the mission achieved 95 percent of its planned objectives.[1][16][17][18] Among its many achievements was the discovery of the widespread presence of water molecules in lunar soil.[19]
Water discovered on moon These images show a very young lunar crater on the side of the moon that faces away from Earth, as viewed by NASA's Moon Mineralogy Mapper on the Indian Space Research Organization's Chandrayaan-1 spacecraft. Credits: ISRO/NASA/JPLCaltech/USGS/Brown Univ.The ISRO's Moon Impact Probe (MIP) onboard Chandrayaan-1 detected water molecules while crash landing on the moon surface.[87] This was confirmed on September 24, 2009, the Science Magazine reported that NASA's Moon Mineralogy Mapper (M3) on Chandrayaan-1 has detected water on the moon.[88] M3 detected absorption features near 2.8-3.0 µm on the surface of the Moon. For silicate bodies, such features are typically attributed to hydroxyl- and/or water-bearing materials. On the Moon, the feature is seen as a widely distributed absorption that appears strongest at cooler high latitudes and at several fresh feldspathic craters. The general lack of correlation of this feature in sunlit M3 data with neutron spectrometer H abundance data suggests that the formation and retention of OH and H2O is an ongoing surficial process. OH/H2O production processes may feed polar cold traps and make the lunar regolith a candidate source of volatiles for human exploration.
The Moon Mineralogy Mapper (M3), an imaging spectrometer, was one of the 11 instruments on board Chandrayaan-I that came to a premature end on August 29. M3 was aimed at providing the first mineral map of the entire lunar surface.
Lunar scientists have for decades contended with the possibility of water repositories. They are now increasingly “confident that the decades-long debate is over,” a report says. “The moon, in fact, has water in all sorts of places; not just locked up in minerals, but scattered throughout the broken-up surface, and, potentially, in blocks or sheets of ice at depth.” The results from the NASA’s Lunar Reconnaissance Orbiter are also “offering a wide array of watery signals.” [89][90]

BARAK OBAMA







Barack Hussein Obama II (/bəˈrɑːk huːˈseɪn oʊˈbɑːmə/ ( listen); born August 4, 1961) is the 44th and current President of the United States. He is the first African American to hold the office, as well as the first born in Hawaii. Obama previously served as the junior United States Senator from Illinois from January 2005 until he resigned after his election to the presidency in November 2008.
Obama is a graduate of Columbia University and Harvard Law School, where he was the president of the Harvard Law Review. He was a community organizer in Chicago before earning his law degree. He worked as a civil rights attorney in Chicago and taught constitutional law at the University of Chicago Law School from 1992 to 2004.
Obama served three terms in the Illinois Senate from 1997 to 2004. Following an unsuccessful bid for a seat in the U.S. House of Representatives in 2000, Obama ran for United States Senate in 2004. His victory, from a crowded field, in the March 2004 Democratic primary raised his visibility. His prime-time televised keynote address at the Democratic National Convention in July 2004 made him a rising star nationally in the Democratic Party. He was elected to the U.S. Senate in November 2004 by the largest margin in the history of Illinois.
He began his run for the presidency in February 2007. After a close campaign in the 2008 Democratic Party presidential primaries against Hillary Rodham Clinton, he won his party's nomination. In the 2008 general election, he defeated Republican nominee John McCain and was inaugurated as president on January 20, 2009.
--------------------------------------------------------------------------------
44th President of the United States Incumbent Assumed office January 20, 2009 Vice President Joe Biden Preceded by George W. Bush
--------------------------------------------------------------------------------
United States Senatorfrom Illinois In officeJanuary 3, 2005 – November 16, 2008 Preceded by Peter Fitzgerald Succeeded by Roland Burris
--------------------------------------------------------------------------------
Member of the Illinois Senatefrom the 13th district In officeJanuary 8, 1997 – November 4, 2004 Preceded by Alice Palmer Succeeded by Kwame Raoul
-------------------------------------------------------------------------------- Born August 4, 1961 (1961-08-04) (age 48)[1]Honolulu, Hawaii[2] Birth name Barack Hussein Obama II[2] Nationality American Political party Democratic Spouse(s) Michelle Obama (m. 1992) Children Malia Ann (b.1998)Natasha (Sasha) (b.2001) Residence The White House (official) Chicago, Illinois (private) Alma mater Occidental CollegeColumbia University (B.A.)Harvard Law School (J.D.) Occupation Community organizerLawyerConstitutional law ProfessorAuthor Religion Christian[3] Obama's family history, early life and upbringing, and Ivy League education differ markedly from those of African-American politicians who launched their careers in the 1960s through participation in the civil rights movement.[197] Expressing puzzlement over questions about whether he is "black enough", Obama told an August 2007 meeting of the National Association of Black Journalists that "we're still locked in this notion that if you appeal to white folks then there must be something wrong."[198] Obama acknowledged his youthful image in an October 2007 campaign speech, saying: "I wouldn't be here if, time and again, the torch had not been passed to a new generation."[199]
Obama presents his first weekly address as President of the United States, discussing the American Recovery and Reinvestment Act of 2009.Obama is frequently referred to as an exceptional orator.[200] During his pre-inauguration transition period and continuing into his presidency, Obama has delivered a series of weekly Internet video addresses[201] similar to Franklin D. Roosevelt's famous fireside chats to explain his policies and actions.[202]
According to the Gallup Daily Poll, during his first 100 days in office as president, Obama received approval ratings in the mid-60s, ranging from 59% to 69%. He concluded his first 100 days with a 65% approval rating.[203] His disapproval rating increased from 12% to 29% during that same time period.[204] By late August 2009, his approval rating had dropped to 50%, with a 42% disapproval rating.[204][205]
Obama's international appeal has been described as a defining factor for his public image.[206] Polls show strong support for Obama in other countries,[207] and he has met with prominent foreign figures including then-British Prime Minister Tony Blair,[208] Italy's Democratic Party leader and then Mayor of Rome Walter Veltroni,[209] and French President Nicolas Sarkozy.[210]
According to a May 2009 poll conducted by Harris Interactive for France 24 and the International Herald Tribune, Obama was rated as the most popular world leader, as well as the one figure most people would pin their hopes on for pulling the world out of this economic downturn.[211]
Obama won Best Spoken Word Album Grammy Awards for abridged audiobook versions of Dreams from My Father in February 2006 and for The Audacity of Hope in February 2008.[212] His "Yes We Can" speech, which artists independently set to music, was viewed by 10 million people on YouTube in the first month,[213] and received a Daytime Emmy Award.[214] In December 2008, Time magazine named Barack Obama as its Person of the Year for his historic candidacy and election, which it described as "the steady march of seemingly impossible accomplishments".[215]

GANDHIJI











Mohandas Karamchand Gandhi (Gujarati: મોહનદાસ કરમચંદ ગાંધી, pronounced [moːɦənˈdaːs kəɾəmˈtʂənd ˈɡaːndʱiː] ( listen); 2 October 1869 – 30 January 1948) was the pre-eminent political and spiritual leader of India during the Indian independence movement. He was the pioneer of satyagraha—resistance to tyranny through mass civil disobedience, firmly founded upon ahimsa or total non-violence—which led India to independence and inspired movements for civil rights and freedom across the world. Gandhi is commonly known around the world as Mahatma Gandhi (Sanskrit: महात्मा mahātmā or "Great Soul", an honorific first applied to him by Rabindranath Tagore),[1] and in India also as Bapu (Gujarati: Gujarati: બાપુ, bāpu or "Father"). He is officially honoured in India as the Father of the Nation; his birthday, 2 October, is commemorated there as Gandhi Jayanti, a national holiday, and worldwide as the International Day of Non-Violence.
Gandhi first employed non-violent civil disobedience while an expatriate lawyer in South Africa, during the resident Indian community's struggle for civil rights. After his return to India in 1915, he organized protests by peasants, farmers, and urban labourers concerning excessive land-tax and discrimination. After assuming leadership of the Indian National Congress in 1921, Gandhi led nationwide campaigns to ease poverty, expand women's rights, build religious and ethnic amity, end untouchability, and increase economic self-reliance. Above all, he aimed to achieve Swaraj or the independence of India from foreign domination. Gandhi famously led his followers in the Non-cooperation movement that protested the British-imposed salt tax with the 400 km (240 mi) Dandi Salt March in 1930. Later he campaigned against the British to Quit India. Gandhi spent a number of years in jail in both South Africa and India.
As a practitioner of ahimsa, he swore to speak the truth and advocated that others do the same. Gandhi lived modestly in a self-sufficient residential community and wore the traditional Indian dhoti and shawl, woven with yarn he had hand spun on a charkha. He ate simple vegetarian food, and also undertook long fasts as a means of both self-purification and social protest.
Mohandas Karamchand Gandhi[2] was born in Porbandar, a coastal town in present-day Gujarat, India, on 2 October 1869. His father, Karamchand Gandhi (1822-1885), who belonged to the Hindu Modh community, was the diwan (Prime Minister) of the eponymous Porbander state, a small princely state in the Kathiawar Agency of British India.[3] His grandfather's name was Uttamchand Gandhi, fondly called Utta Gandhi. His mother, Putlibai, who came from the Hindu Pranami Vaishnava community, was Karamchand's fourth wife, the first three wives having apparently died in childbirth.[4] Growing up with a devout mother and the Jain traditions of the region, the young Mohandas absorbed early the influences that would play an important role in his adult life; these included compassion for sentient beings, vegetarianism, fasting for self-purification, and mutual tolerance between individuals of different creeds.[citation needed]
The Indian classics, especially the stories of Shravana and Maharaja Harishchandra from the Indian epics, had a great impact on Gandhi in his childhood. The story of Harishchandra, a well known tale of an ancient Indian king and a truthful hero, haunted Gandhi as a boy. Gandhi in his autobiography admits that it left an indelible impression on his mind. He writes: "It haunted me and I must have acted Harishchandra to myself times without number." Gandhi's early self-identification with Truth and Love as supreme values is traceable to his identification with these epic characters.[5][6]
In May 1883, the 13-year old Mohandas was married to 14-year old Kasturbai Makhanji (her first name was usually shortened to "Kasturba," and affectionately to "Ba") in an arranged child marriage, as was the custom in the region.[7] Recalling about the day of their marriage he once said that " As we didn't know much about marriage, for us it meant only wearing new clothes, eating sweets and playing with relatives." However, as was also the custom of the region, the adolescent bride was to spend much time at her parents' house, and away from her husband.[8] In 1885, when Gandhi was 15, the couple's first child was born, but survived only a few days; Gandhi's father, Karamchand Gandhi, had died earlier that year.[9] Mohandas and Kasturba had four more children, all sons: Harilal, born in 1888; Manilal, born in 1892; Ramdas, born in 1897; and Devdas, born in 1900. At his middle school in Porbandar and high school in Rajkot, Gandhi remained an average student academically. He passed the matriculation exam for Samaldas College at Bhavnagar, Gujarat with some difficulty. While there, he was unhappy, in part because his family wanted him to become a barrister.
Gandhi and his wife Kasturba (1902)On 4 September 1888, less than a month shy of his 19th birthday, Gandhi traveled to London, England, to study law at University College London and to train as a barrister. His time in London, the Imperial capital, was influenced by a vow he had made to his mother in the presence of the Jain monk Becharji, upon leaving India, to observe the Hindu precepts of abstinence from meat, alcohol, and promiscuity.[10] Although Gandhi experimented with adopting "English" customs—taking dancing lessons for example—he could not stomach the bland vegetarian food offered by his landlady and he was always hungry until he found one of London's few vegetarian restaurants. Influenced by Salt's book, he joined the Vegetarian Society, was elected to its executive committee[10], and started a local Bayswater chapter.[4] Some of the vegetarians he met were members of the Theosophical Society, which had been founded in 1875 to further universal brotherhood, and which was devoted to the study of Buddhist and Hindu literature. They encouraged Gandhi to join them in reading the Bhagavad Gita both in translation as well as in the original.[10] Not having shown a particular interest in religion before, he became interested in religious thought and began to read both Hindu as well as Christian scriptures.[4][10]
Gandhi was called to the bar on June 10, 1891 and left London for India on June 12, 1891,[4] where he learned that his mother had died while he was in London, his family having kept the news from him.[10] His attempts at establishing a law practice in Mumbai failed and, later, after applying and being turned down for a part-time job as a high school teacher, he ended up returning to Rajkot to make a modest living drafting petitions for litigants, a business he was forced to close when he ran afoul of a British officer. In his autobiography, he refers to this incident as an unsuccessful attempt to lobby on behalf of his older brother.[4][10] It was in this climate that, in April 1893, he accepted a year-long contract from Dada Abdulla & Co., an Indian firm, to a post in the Colony of Natal, South Africa, then part of the British Empire.[4]
Gandhi advised the Congress to reject the proposals the British Cabinet Mission offered in 1946, as he was deeply suspicious of the grouping proposed for Muslim-majority states—Gandhi viewed this as a precursor to partition. However, this became one of the few times the Congress broke from Gandhi's advice (though not his leadership), as Nehru and Patel knew that if the Congress did not approve the plan, the control of government would pass to the Muslim League. Between 1946 and 1948, over 5,000 people were killed in violence. Gandhi was vehemently opposed to any plan that partitioned India into two separate countries. An overwhelming majority of Muslims living in India, side by side with Hindus and Sikhs, were in favour of Partition. Additionally Muhammad Ali Jinnah, the leader of the Muslim League, commanded widespread support in West Punjab, Sindh, North-West Frontier Province and East Bengal. The partition plan was approved by the Congress leadership as the only way to prevent a wide-scale Hindu-Muslim civil war. Congress leaders knew that Gandhi would viscerally oppose partition, and it was impossible for the Congress to go ahead without his agreement, for Gandhi's support in the party and throughout India was strong. Gandhi's closest colleagues had accepted partition as the best way out, and Sardar Patel endeavoured to convince Gandhi that it was the only way to avoid civil war. A devastated Gandhi gave his assent.
He conducted extensive dialogue with Muslim and Hindu community leaders, working to cool passions in northern India, as well as in Bengal. Despite the Indo-Pakistani War of 1947, he was troubled when the Government decided to deny Pakistan the 55 crores (550 million Indian rupees) due as per agreements made by the Partition Council. Leaders like Sardar Patel feared that Pakistan would use the money to bankroll the war against India. Gandhi was also devastated when demands resurged for all Muslims to be deported to Pakistan, and when Muslim and Hindu leaders expressed frustration and an inability to come to terms with one another.[29]. Gandhi's arrival in Delhi, turned out to an important intervention in ending the rioting, he even visited Muslims mohallas to restore faith of the Muslim populace. He launched his last fast-unto-death on January 12, 1948, in Delhi [30], asking that all communal violence be ended once and for all, Muslims homes be restored to them and that the payment of 550 million rupees be made to Pakistan. Gandhi feared that instability and insecurity in Pakistan would increase their anger against India, and violence would spread across the borders. He further feared that Hindus and Muslims would renew their enmity and that this would precipitate open civil war. After emotional debates with his life-long colleagues, Gandhi refused to budge, and the Government rescinded its policy and made the payment to Pakistan. Hindu, Muslim and Sikh community leaders, including the Rashtriya Swayamsevak Sangh and Hindu Mahasabha assured him that they would renounce violence and call for peace. Gandhi thus broke his fast by sipping orange juice.[31]
Mohandas Karamchand Gandhi in Bombay, 1944. Born October 2, 1869(1869-10-02)Porbandar, Kathiawar Agency, British India Died January 30, 1948 (aged 78)New Delhi, Union of India Cause of death Assassination Resting place Rajghat in New Delhi Nationality Indian Other names Mahatma Gandhi, Bapu Alma mater University College London Known for Prominent Figure of Indian Independence MovementPropounding the philosophy of Satyagraha and Ahimsa Religious beliefs Hinduism Spouse(s) Kasturba Gandhi Children HarilalManilalRamdasDevdas Parents Putlibai Gandhi (Mother)Karamchand Gandhi (Father) Signature

MICHEAL JACSON







Michael Joseph Jackson (August 29, 1958 – June 25, 2009), known as the "King of Pop", was an American musician and one of the most commercially successful and influential entertainers of all time. His unique contributions to music and dance, along with a highly publicized personal life, made him a prominent figure in popular culture for four decades.
He made his debut in 1964 with his brothers as a member of the Motown group The Jackson 5; singing hits such as "I Want You Back" and "ABC" and started a solo career in 1971. His 1982 album Thriller remains the best-selling album of all time, with Off the Wall (1979), Bad (1987), Dangerous (1991), and HIStory (1995) among the best selling.
Along with song success, he is widely credited with having transformed the music video from a promotional tool into an art form, with videos for his songs "Billie Jean", "Beat It" and "Thriller" making him the first African American artist to amass a strong crossover following on MTV - helping bring the relatively new channel to fame. Videos such as "Black or White" and "Scream" kept Jackson as a staple on MTV into the 1990s. With stage performances and music videos, Jackson popularized a number of physically complicated dance techniques, such as the robot and the moonwalk, embedding them into popular culture. His distinctive musical sound, vocal style and dance has inspired numerous hip hop, pop and contemporary R&B artists, while also breaking down cultural, racial and generational barriers.
One of the few artists to have been inducted into the Rock and Roll Hall of Fame twice, his other achievements feature multiple Guinness World Records—including the "Most Successful Entertainer of All Time"—13 Grammy Awards, 17 number one singles (including the four as a member of the Jackson 5), and estimated sales between 350 million and 750 million records worldwide;[1] making him one of the top three best selling artists in history. He was also a notable philanthropist and humanitarian who donated and raised millions of dollars through support of 39 charities and his own Heal the World Foundation.
Jackson's personal life generated controversy for years. His changing appearance was noticed from the late 1970s onwards, with changes to his nose and to the color of his skin drawing media publicity. He was accused of child sexual abuse in 1993 though no charges were brought, and in 2005 he was tried and acquitted of second allegations. He married twice, first in 1994 and again in 1996, and brought up three children, one born to a surrogate mother. While preparing for the This Is It concert tour in 2009, Jackson died at the age of 50 after suffering from cardiac arrest. He reportedly had been administered drugs such as propofol and lorazepam, and his death was ruled a homicide by the Los Angeles County coroner. His memorial service was broadcast live around the world, attracting a global audience of up to one billion people.[2]
Birth name Michael Joseph Jackson Born August 29, 1958(1958-08-29)Gary, Indiana, United States Died June 25, 2009 (aged 50)Los Angeles, California,United States Genres Rhythm and blues, pop, rock, dance Occupations Singer-songwriter, record producer, dancer, choreographer, actor, businessman Instruments Vocals, multiple instruments Years active 1964–2009 Labels Motown, Epic Associated acts The Jackson 5 Website www.michaeljackson.com/



Sunday, October 4, 2009

HISTORY OF CHEMISTRY

By 1000 BC, the ancient civilizations were using technologies that would form the basis of the various branches of chemistry. Extracting metal from their ores, making pottery and glazes, fermenting beer and wine, making pigments for cosmetics and painting, extracting chemicals from plants for medicine and perfume, making cheese, dying cloth, tanning leather, rendering fat into soap, making glass, and making alloys like bronze.
Philosophical attempts to explain the nature of matter and its transformations failed. The protoscience of alchemy also failed, but by experimentation and recording the results set the stage for science. Modern chemistry begins to emerge when a clear distinction is made between chemistry and alchemy by Robert Boyle in his work The Sceptical Chymist (1661). Chemistry then becomes a full-fledged science when Antoine Lavoisier develops his law of conservation of mass, which demands careful measurements and quantitative observations of chemical phenomena. So, while both alchemy and chemistry are concerned with the nature of matter and its transformations, it is only the chemists who apply the scientific method. The history of chemistry is intertwined with the history of thermodynamics, especially through the work of Willard Gibbs.[1]
From fire to atomism
Arguably the first chemical reaction that was used in a controlled manner by mankind was fire. However, for millennia, in absence of a scientific understanding, fire was simply a mystical force that could transform one substance into another (burn the wood, or boil the water) while producing heat and light. Fire affected many aspects of early societies, ranging from the most simple facets of everyday life, such as cooking and habitat lighting, to more advanced technologies, such as pottery, bricks, and smelting of metals to make tools.
Philosophical attempts to rationalize why different substances have different properties (color, density, smell), exist in different states (gaseous, liquid, and solid), and react in a different manner when exposed to environments, for example to water or fire or temperature changes, led ancient philosophers to postulate the first theories on nature and chemistry. The history of such philosophical theories that relate to chemistry, can probably be traced back to every single ancient civilization. The common aspect in all these theories was the attempt to identify a small number of primary elements that make up all the various substances in nature. Substances like air, water, and soil/earth, energy forms, such as fire and light, and more abstract concepts such as ideas, aether, and heaven, were common in ancient civilizations even in absence of any cross-fertilization; for example in Greek, Indian, Mayan, and ancient Chinese philosophies all considered air, water, earth and fire as primary elements.[citation needed]
Atomism can be traced back to ancient Greece and ancient India.[2] Greek atomism dates back to 440 BC, as what might be indicated by the book De Rerum Natura (The Nature of Things)[3] written by the Roman Lucretius[4] in 50 BC. In the book was found ideas traced back to Democritus and Leucippus, who declared that atoms were the most indivisible part of matter. This coincided with a similar declaration by Indian philosopher Kanada in his Vaisheshika sutras around the same time period.[2] By similar means discussed the existence of gases. What Kanada declared by sutra, Democritus declared by philosophical musing. Both suffered from a lack of empirical data. Without scientific proof, the existence of atoms was easy to deny. Aristotle opposed the existence of atoms in 330 BC; and the atomism of the Vaisheshika school was also opposed for a long time.[citation needed]
Much of the early development of purification methods is described by Pliny the Elder in his Naturalis Historia. He made attempts to explain those methods, as well as making acute observations of the state of many minerals.
[edit] The rise of metallurgy
Main articles: History of ferrous metallurgy and History of metallurgy in the Indian subcontinent
It was fire that led to the discovery of glass and the purification of metals which in turn gave way to the rise of metallurgy.[citation needed] During the early stages of metallurgy, methods of purification of metals were sought, and gold, known in ancient Egypt as early as 2600 BC, became a precious metal. The discovery of alloys heralded the Bronze Age. After the Bronze Age, the history of metallurgy was marked by which army had better weaponry. Countries in Eurasia had their heyday when they made the superior alloys, which, in turn, made better armour and better weapons. This often determined the outcomes of battles.[citation needed] Significant progress in metallurgy and alchemy was made in ancient India.[5]
[edit] The philosopher's stone and the rise of alchemy
Main article: Alchemy

"Renel the Alchemist", by Sir William Douglas, 1853
Many people were interested in finding a method that could convert cheaper metals into gold. The material that would help them do this was rumored to exist in what was called the philosopher's stone. This led to the protoscience called alchemy. Alchemy was practiced by many cultures throughout history and often contained a mixture of philosophy, mysticism, and protoscience.[citation needed]
Alchemy not only sought to turn base metals into gold, but especially in a Europe rocked by bubonic plague, there was hope that alchemy would lead to the development of medicines to improve people's health. The holy grail of this strain of alchemy was in the attempts made at finding the elixir of life, which promised eternal youth. Neither the elixir nor the philosopher's stone were ever found. Also, characteristic of alchemists was the belief that there was in the air an "ether" which breathed life into living things.[citation needed] Practitioners of alchemy included Isaac Newton, who remained one throughout his life.
[edit] Problems encountered with alchemy
There were several problems with alchemy, as seen from today's standpoint. There was no systematic naming system for new compounds, and the language was esoteric and vague to the point that the terminologies meant different things to different people. In fact, according to The Fontana History of Chemistry (Brock, 1992):
The language of alchemy soon developed an arcane and secretive technical vocabulary designed to conceal information from the uninitiated. To a large degree, this language is incomprehensible to us today, though it is apparent that readers of Geoffery Chaucer's Canon's Yeoman's Tale or audiences of Ben Jonson's The Alchemist were able to construe it sufficiently to laugh at it.[6]
Chaucer's tale exposed the more fraudulent side of alchemy, especially the manufacture of counterfeit gold from cheap substances. Soon after Chaucer, Dante Alighieri also demonstrated an awareness of this fraudulence, causing him to consign all alchemists to the Inferno in his writings. Soon after, in 1317, the Avignon Pope John XXII ordered all alchemists to leave France for making counterfeit money. A law was passed in England in 1403 which made the "multiplication of metals" punishable by death. Despite these and other apparently extreme measures, alchemy did not die. Royalty and privileged classes still sought to discover the philosopher's stone and the elixir of life for themselves.[7]
There was also no agreed-upon scientific method for making experiments reproducible. Indeed many alchemists included in their methods irrelevant information such as the timing of the tides or the phases of the moon. The esoteric nature and codified vocabulary of alchemy appeared to be more useful in concealing the fact that they could not be sure of very much at all. As early as the 14th century, cracks seemed to grow in the facade of alchemy; and people became sceptical.[citation needed] Clearly, there needed to be a scientific method where experiments can be repeated by other people, and results needed to be reported in a clear language that laid out both what is known and unknown.
[edit] From alchemy to chemistry
[edit] Early chemists
See also: Timeline of chemistry and Alchemy and chemistry in Islam

Jabir ibn Hayyan (Geber), an Arabic alchemist whose experimental research laid the foundations for chemistry.
In the Arab world, the Muslims were translating the works of the ancient Greeks into Arabic and were experimenting with scientific ideas.[8] The development of the modern scientific method was slow and arduous, but an early scientific method for chemistry began emerging among early Muslim chemists, beginning with the 9th century chemist Jabir ibn Hayyan (known as "Geber" in Europe), who is "considered by many to be the father of chemistry".[9][10][11][12] He introduced a systematic and experimental approach to scientific research based in the laboratory, in contrast to the ancient Greek and Egyptian alchemists whose works were often allegorical and unintelligble.[13] He also invented and named the alembic (al-anbiq), chemically analyzed many chemical substances, composed lapidaries, distinguished between alkalis and acids, and manufactured hundreds of drugs.[14]
Among other influential Muslim chemists, Ja'far al-Sadiq,[15] Alkindus,[16] Abū al-Rayhān al-Bīrūnī,[17] Avicenna[18] and Ibn Khaldun refuted the practice of alchemy and the theory of the transmutation of metals; and Tusi described an early version of the conservation of mass, noting that a body of matter is able to change but is not able to disappear.[19] Rhazes refuted Aristotle's theory of four classical elements for the first time and set up the firm foundations of modern chemistry, using the laboratory in the modern sense, designing and describing more than twenty instruments, many parts of which are still in use today, such as a crucible, decensory, cucurbit or retort for distillation, and the head of a still with a delivery tube (ambiq, Latin alembic), and various types of furnace or stove.[20]

Agricola, author of De re metallica
For the more honest practitioners in Europe, alchemy became an intellectual pursuit after early Arabic alchemy became available through Latin translation, and over time, they got better at it. Paracelsus (1493-1541), for example, rejected the 4-elemental theory and with only a vague understanding of his chemicals and medicines, formed a hybrid of alchemy and science in what was to be called iatrochemistry. Paracelsus was not perfect in making his experiments truly scientific. For example, as an extension of his theory that new compounds could be made by combining mercury with sulfur, he once made what he thought was "oil of sulfur". This was actually dimethyl ether, which had neither mercury nor sulfur.[citation needed]
Practical attempts to improve the refining of ores and their extraction to smelt metals was an important source of information for early chemists, among them Georg Agricola (1494–1555), who published his great work De re metallica in 1556. His approach removed the mysticism associated with the subject, creating the practical base upon which others could build. The work describes the many kinds of furnace used to smelt ore, and stimulated interest in minerals and their composition. It is no coincidence that he gives numerous references to the earlier author, Pliny the Elder and his Naturalis Historia.
In 1605, Sir Francis Bacon published The Proficience and Advancement of Learning, which contains a description of what would later be known as the scientific method.[21] In 1615 Jean Beguin publishes the Tyrocinium Chymicum, an early chemistry textbook, and in it draws the first-ever chemical equation.[22]

Robert Boyle, one of the co-founders of modern chemistry through his use of proper experimentation, which further separated chemistry from alchemy
Robert Boyle (1627–1691) is considered to have refined the modern scientific method for alchemy and to have separated chemistry further from alchemy.[23] Robert Boyle was an atomist, but favoured the word corpuscle over atoms. He comments that the finest division of matter where the properties are retained is at the level of corpuscles.
Boyle was credited with the discovery of Boyle's Law. He is also credited for his landmark publication The Sceptical Chymist, where he attempts to develop an atomic theory of matter, with no small degree of success. Despite all these advances, the person celebrated as the "father of modern chemistry" is Antoine Lavoisier who developed his law of conservation of mass in 1789, also called Lavoisier's Law.[24] With this, chemistry acquired a strict quantitative nature, allowing reliable predictions to be made.
In 1754, Joseph Black isolated carbon dioxide, which he called "fixed air".[25] Carl Wilhelm Scheele and Joseph Priestly independently isolated oxygen, called by Priestly "dephlogisticated air" and Scheele "fire air".[26][27]
Joseph Proust proposed the law of definite proportions, which states that elements always combine in small, whole number ratios to form compounds.[28] In 1800, Alessandro Volta devised the first chemical battery, thereby founding the discipline of electrochemistry.[29] In 1803, John Dalton proposed Dalton's Law, which describes relationship between the components in a mixture of gases and the relative pressure each contributes to that of the overall mixture.[30]
[edit] Antoine Lavoisier

Portrait of Monsieur Lavoisier and his wife, by Jacques-Louis David
Although the archives of chemical research draw upon work from ancient Babylonia, Egypt, and especially the Arabs and Persians after Islam, modern chemistry flourished from the time of Antoine Lavoisier, who is regarded as the "father of modern chemistry", particularly for his discovery of the law of conservation of mass, and his refutation of the phlogiston theory of combustion in 1783. (Phlogiston was supposed to be an imponderable substance liberated by flammable materials in burning.) Mikhail Lomonosov independently established a tradition of chemistry in Russia in the 18th century.[citation needed] Lomonosov also rejected the phlogiston theory, and anticipated the kinetic theory of gases.[citation needed] He regarded heat as a form of motion, and stated the idea of conservation of matter.
[edit] The vitalism debate and organic chemistry
After the nature of combustion (see oxygen) was settled, another dispute, about vitalism and the essential distinction between organic and inorganic substances, was revolutionized by Friedrich Wöhler's accidental synthesis of urea from inorganic substances in 1828. Never before had an organic compound been synthesized from inorganic material.[citation needed] This opened a new research field in chemistry, and by the end of the 19th century, scientists were able to synthesize hundreds of organic compounds. The most important among them are mauve, magenta, and other synthetic dyes, as well as the widely used drug aspirin. The discovery of the artificial synthesis of urea contributed greatly to the theory of isomerism, as the empirical chemical formulas for urea and ammonium cyanate can be expressed similarly.
[edit] Disputes about atomism after Lavoisier

Bust of John Dalton by Chantrey
Throughout the 19th century, chemistry was divided between those who followed the atomic theory of John Dalton and those who did not, such as Wilhelm Ostwald and Ernst Mach.[31] Although such proponents of the atomic theory as Amedeo Avogadro and Ludwig Boltzmann made great advances in explaining the behavior of gases, this dispute was not finally settled until Jean Perrin's experimental investigation of Einstein's atomic explanation of Brownian motion in the first decade of the 20th century.[31]
Well before the dispute had been settled, many had already applied the concept of atomism to chemistry. A major example was the ion theory of Svante Arrhenius which anticipated ideas about atomic substructure that did not fully develop until the 20th century. Michael Faraday was another early worker, whose major contribution to chemistry was electrochemistry, in which (among other things) a certain quantity of electricity during electrolysis or electrodeposition of metals was shown to be associated with certain quantities of chemical elements, and fixed quantities of the elements therefore with each other, in specific ratios.[citation needed] These findings, like those of Dalton's combining ratios, were early clues to the atomic nature of matter.
[edit] The periodic table
Main article: History of the periodic table

Dmitri Mendeleev, responsible for the periodic table.
For many decades, the list of known chemical elements had been steadily increasing. A great breakthrough in making sense of this long list (as well as in understanding the internal structure of atoms as discussed below) was Dmitri Mendeleev and Lothar Meyer's development of the periodic table, and particularly Mendeleev's use of it to predict the existence and the properties of germanium, gallium, and scandium, which Mendeleev called ekasilicon, ekaaluminium, and ekaboron respectively. Mendeleev made his prediction in 1870; gallium was discovered in 1875, and was found to have roughly the same properties that Mendeleev predicted for it.[citation needed]
[edit] The modern definition of chemistry
Classically, before the 20th century, chemistry was defined as the science of the nature of matter and its transformations. It was therefore clearly distinct from physics which was not concerned with such dramatic transformation of matter. Moreover, in contrast to physics, chemistry was not using much of mathematics. Even some were particularly reluctant to using mathematics within chemistry. For example, Auguste Comte wrote in 1830:
Every attempt to employ mathematical methods in the study of chemical questions must be considered profoundly irrational and contrary to the spirit of chemistry.... if mathematical analysis should ever hold a prominent place in chemistry -- an aberration which is happily almost impossible -- it would occasion a rapid and widespread degeneration of that science.
However, in the second part of the 19th century, the situation changed and August Kekule wrote in 1867:
I rather expect that we shall someday find a mathematico-mechanical explanation for what we now call atoms which will render an account of their properties.
After the discovery by Ernest Rutherford and Niels Bohr of the atomic structure in 1912, and by Marie and Pierre Curie of radioactivity, scientists had to change their viewpoint on the nature of matter. The experience acquired by chemists was no longer pertinent to the study of the whole nature of matter but only to aspects related to the electron cloud surrounding the atomic nuclei and the movement of the latter in the electric field induced by the former (see Born-Oppenheimer approximation). The range of chemistry was thus restricted to the nature of matter around us in conditions which are not too far from standard conditions for temperature and pressure and in cases where the exposure to radiation is not too different from the natural microwave, visible or UV radiations on Earth. Chemistry was therefore re-defined as the science of matter that deals with the composition, structure, and properties of substances and with the transformations that they undergo.[citation needed] However the meaning of matter used here relates explicitly to substances made of atoms and molecules, disregarding the matter within the atomic nuclei and its nuclear reaction or matter within highly ionized plasmas. Nevertheless the field of chemistry is still, on our human scale, very broad and the claim that chemistry is everywhere is accurate.
[edit] Quantum chemistry
Main article: Quantum chemistry
Some view the birth of quantum chemistry in the discovery of the Schrödinger equation and its application to the hydrogen atom in 1926.[citation needed] However, the 1927 article of Walter Heitler and Fritz London[32] is often recognised as the first milestone in the history of quantum chemistry.[33] This is the first application of quantum mechanics to the diatomic hydrogen molecule, and thus to the phenomenon of the chemical bond. In the following years much progress was accomplished by Edward Teller, Robert S. Mulliken, Max Born, J. Robert Oppenheimer, Linus Pauling, Erich Hückel, Douglas Hartree, Vladimir Aleksandrovich Fock, to cite a few.[citation needed]
Still, skepticism remained as to the general power of quantum mechanics applied to complex chemical systems.[citation needed] The situation around 1930 is described by Paul Dirac:[34]
The underlying physical laws necessary for the mathematical theory of a large part of physics and the whole of chemistry are thus completely known, and the difficulty is only that the exact application of these laws leads to equations much too complicated to be soluble. It therefore becomes desirable that approximate practical methods of applying quantum mechanics should be developed, which can lead to an explanation of the main features of complex atomic systems without too much computation.
Hence the quantum mechanical methods developed in the 1930s and 1940s are often referred to as theoretical molecular or atomic physics to underline the fact that they were more the application of quantum mechanics to chemistry and spectroscopy than answers to chemically relevant questions.
In the 1940s many physicists turned from molecular or atomic physics to nuclear physics (like J. Robert Oppenheimer or Edward Teller). In 1951, a milestone article in quantum chemistry is the seminal paper of Clemens C. J. Roothaan on Roothaan equations.[35] It opened the avenue to the solution of the self-consistent field equations for small molecules like hydrogen or nitrogen. Those computations were performed with the help of tables of integrals which were computed on the most advanced computers of the time.[citation needed]
[edit] Molecular biology and biochemistry
Main articles: History of molecular biology and History of biochemistry
By the mid 20th century, in principle, the integration of physics and chemistry was extensive, with chemical properties explained as the result of the electronic structure of the atom; Linus Pauling's book on The Nature of the Chemical Bond used the principles of quantum mechanics to deduce bond angles in ever-more complicated molecules. However, though some principles deduced from quantum mechanics were able to predict qualitatively some chemical features for biologically relevant molecules, they were, till the end of the 20th century, more a collection of rules, observations, and recipes than rigorous ab initio quantitative methods.[citation needed]

Diagrammatic representation of some key structural features of DNA
This heuristic approach triumphed in 1953 when James Watson and Francis Crick deduced the double helical structure of DNA by constructing models constrained by and informed by the knowledge of the chemistry of the constituent parts and the X-ray diffraction patterns obtained by Rosalind Franklin.[36] This discovery lead to an explosion of research into the biochemistry of life.
In the same year, the Miller-Urey experiment demonstrated that basic constituents of protein, simple amino acids, could themselves be built up from simpler molecules in a simulation of primordial processes on Earth. Though many questions remain about the true nature of the origin of life, this was the first attempt by chemists to study hypothetical processes in the laboratory under controlled conditions.[citation needed]
In 1983 Kary Mullis devised a method for the in-vitro amplification of DNA, known as the polymerase chain reaction (PCR), which revolutionized the chemical processes used in the laboratory to manipulate it. PCR could be used to synthesize specific pieces of DNA and made possible the sequencing of DNA of organisms, which culminated in the huge human genome project.
An important piece in the double helix puzzle was solved by one of Pauling's student Matthew Meselson and Frank Stahl, the result of their collaboration (Meselson-Stahl experiment) has been called as "the most beatiful experiment in biology".
They used a centrifugation technique that sorted molecules according to differences in weight. Because nitrogen atoms are a component of DNA, they were labelled and therefore tracked in replication in bacteria.
[edit] Chemical industry
Main article: Chemical industry
The later part of the nineteenth century saw a huge increase in the exploitation of petroleum extracted from the earth for the production of a host of chemicals and largely replaced the use of whale oil, coal tar and naval stores used previously. Large scale production and refinement of petroleum provided feedstocks for liquid fuels such as gasoline and diesel, solvents, lubricants, asphalt, waxes, and for the production of many of the common materials of the modern world, such as synthetic fibers, plastics, paints, detergents, pharmaceuticals, adhesives and ammonia as fertilizer and for other uses. Many of these required new catalysts and the utilization of chemical engineering for their cost-effective production.
In the mid-twentieth century, control of the electronic structure of semiconductor materials was made precise by the creation of large ingots of extremely pure single crystals of silicon and germanium. Accurate control of their chemical composition by doping with other elements made the production of the solid state transistor in 1951 and made possible the production of tiny integrated circuits for use in electronic devices, especially computers.

GLOBAL WARMING

Global climate change impacts in the United States are spelled out with renewed authority in a report released June 16 by the federal government.
The report's key information has been well reported here and in Earth Under Fire and other books, but bears repeating in its straightforward language and up-to-date numbers.
Human activities have led to large increases in heat-trapping gases over the past century. The global warming of the past 50 years is due primarily to this human-induced increase. Global average temperature and sea level have increased, and precipitation patterns have changed.
Human “fingerprints” also have been identified in many other aspects of the climate system, including changes in ocean heat content, precipitation, atmospheric moisture, plant and animal health and location, and Arctic sea ice.
In the U.S., the amount of rain falling in the heaviest downpours has increased approximately 20 percent on average in the past century.
Many types of extreme weather events, such as heat waves and regional droughts, have become more frequent and intense during the past 40 to 50 years. The destructive energy of Atlantic hurricanes has increased... In the eastern Pacific, the strongest hurricanes have become stronger since the 1980s, even while the total number of storms has decreased.
Sea level has risen along most of the U.S. coast over the last 50 years, and will rise more in the future. Arctic sea ice is declining rapidly and this is very likely to continue. Global temperatures are projected to continue to rise over this century.
Whether by 2-3 degrees F or more than 11 degrees depends on a number of factors, including the amount of heat-trapping gas emissions humans continue to allow and how sensitive the climate is to those emissions. Lower emissions of heat-trapping gases will delay the appearance of climate change impacts andlessen their magnitude.
Unless the rate of emissions is substantially reduced, impacts are expected to become increasingly severe for more people and places.
For more from this report, please go to the Temperate Zone page.
Obama's Climate Team Moves to Regulate Greenhouse Gases as Research Shows Global Warming Continues at High Rates
President Barack Obama's new Environmental Protection Agency chief Lisa Jackson has moved to put CO2 and other greenhouse gases under regulation by the Clean Air Act. In one of the most anticipated early actions by the new Administration, the EPA issued a proposed finding on April 17 that these gases endanger human health and well-being. When made final, this will clear the way for regulation of vehicle exhaust, which is the source of about 30 percent of US carbon dioxide emissions.
This is one of the most visible of the climate actions springing from members of the President's new Cabinet, which includes leading scientists and informed diplomats. As they took their posts, working scientists announced in two international meetings that many factors in rapid global warming were getting worse or running at rates which only a few years ago were thought to be extreme.
Besides Jackson, who an was an experienced state environment leader before taking over at EPA, Obama appointed former EPA head Carol Browner to a new post of White House climate and energy chief; Nobel Prize winner Stephen Chu as Secretary of Energy; Harvard professor John Holdren, who has been outspoken on the dangers of climate disruption, as Presidential science advisor; and acclaimed ocean scientist Jane Lubchenco as head of NOAA.
Secretary of State Hillary Clinton replaced George Bush's footdragging international climate negotiators with a team lead by Todd Stern. One of his first actions was to announce to international climate talks in Bonn that "the science is clear, and the threat is real. The facts on the ground are outstripping the worst case scenarios. The costs of inaction-or inadequate actions-are unacceptable." The Bonn talks are preliminary to crucial UN Climate Convention meetings in Copenhagen in December [[link: http://unfccc.int/2860.php]], at which nations have promised to agree to sharp limits on greenhouse gases, replacing the Kyoto Protocol. Many national issues and roadblocks remain, however, prime of which is the world recession which dominates other international meetings.
The EPA finding, although initially focused directly on vehicle emissions, will lead under the Clean Air Act to regulation of greenhouse gas emissions from power plants, source of nearly half of American CO2. Congress is also proposing control of emissions using a cap and trade process familiar to many from previous Clean Air Act procedures to limit sulphur pollution from coal burning plants. A comprehensive climate and energy bill, drafted by Rep. Henry Waxman of California and Rep. Edward Markey of Massachusetts, will be debated in the House this spring. Reactions to the proposed legislation are being posted by many business and environmental groups and will surely intensify as the bill is amended and moves toward a vote later this year.
The urgency of climate action is even greater now because some recent observations are at or beyond the highest projections of previous reports. Scientific studies updating the IPCC assessment of 2007 show that more CO2 is being put into the air than ever before. Rates of change of global mean temperature, sea level rise, ice sheet changes in Greenland and the edges of Antarctica, and ocean chemical changes are running at the highest projections of the 2007 IPCC. In February 2009 at the annual meeting of the American Association for the Advancement of Science, Dr. Chris Field of the Carnegie Institute also reported that some major ways that the earth naturally absorbs CO2 were less efficient now, leaving more of the gas in the air. I heard him say that because of all this, we are "on a trajectory of climate... that has not been explored."
Not every indication of climate is changing this rapidly, but most scientists now predict a 5 degree F or more temperature increase and at least three feet of sea level rise before 2100 if things continue in this way. The changes documented in these website pages and my book occurred during a time of just over one degree of warming.
Every citizen of the world needs to be aware of rapid climate change:
1. Understand the problem, its causes and threats.2. Let your leaders know the facts and that you expect them to act.3. Do something today to reduce greenhouse gas output - please Take action.

EID MUBARAK

Bhuvan mapping

DOWN LOAD Bhuvan mapping A review of ISRO Bhuvan Features and Performance

Here is a frank review of the features and performance of ISRO Bhuvan (the much anticipated satellite-based 3D mapping application from ISRO) BETA Release and comparing it to supposed arch rival Google Earth. Bhuvan from the begining is claiming that it is not competing with Google Earth in any way, but there was much hype and propaganda in the media saying that ISRO Bhuvan will be a Google Earth killer atleast in India. But it looks like that can nit be the case anytime soon. Here is why..

  • While Google Earth works on a downloadable client, Bhuvan works within the browser (only supports Windows and IE 6 and above).
  • The ISRO Bhuvan currently has serious performance issues. The site currently very unstable. It gives up or hangs the browser every once in a while. When a layer (state, district, taluk, etc.) is turned on, it renders unevenly and sometimes fails to render at all. The navigation panel failed to load routinely and it felt like a rare sighting when we could actually use the panel.
  • The promise of high resolution images has not been kept. While the service promises zoom up to 10 metres from the ground level as against 200 metres for Google Earth, we didn’t encounter a single image with nearly as much detailing. In fact, comparative results for a marquee location such as New Delhi’s Connaught Place or Red Fort make its clear as to the inferior performance of ISRO Earth as of now.
  • The navigation tools are similar to Google Earth (GE).
  • The search doesn’t work if a query returns multiple results. A pop up window is supposed to give the multiple results from which the user is supposed to be able to choose. During two days of sporadic testing, we found the result only once. The rest of the time, the window would pop up, but nothing would be displayed. When the search is accurate, the software ‘flies in’ to the exact location, the same way as GE.
  • Users need to create an account and download a plug-in.
  • Bhuvan packs a lot of data on weather, waterbodies and population details of various administrative units. We were unable to access weather data. Clicking on icons of administrative units show basic information such as the population. For specialist users, Bhuvan might hold some attraction. For instance, there is a drought map which cab be used to compare drought situation across years and there is a flood map that shows Bihar during the Kosi flood and after. With Isro backing, Bhuvan would be able to provide such relevant data from time to time, but the application needs major improvements in terms of usability before it will be of interest to the ordinary user.
  • Users can also not edit any data or tag locations.
  • We hope Bhuvan is able to fix the bugs soon. But even then, to be a credible alternative to existing mapping services, and even to get new users to try it, it much provide much higher resolution images. User interest will be piqued only when they can see their house or school or local street in high resolution. With Isro data, this is easily doable.

Having said all this, ISRO Bhuvan is still a very good step forward for ISRO in the right direction we feel. We wish all the best for ISRO and hopefully Bhuvan will mature very fast to become a good service and can really compete with Google Earth.

download bhuvan mapping.............. here