future timeline technology singularity humanity
 
   
future timeline twitter future timeline facebook group future timeline youtube channel account videos future timeline rss feed
 
     
     
 
       
 
 
 

Blog » Energy & the Environment

 
     
 

21st October 2014

2014 on track for hottest year ever

Globally, 2014 is on track for the hottest year ever. September 2014 was the hottest September on record, after the hottest August, which was part of the hottest summer on record. The past 12 months — October 2013–September 2014 — were the warmest 12-month period among all months since records began in 1880.

 

2014 global temperature records

 

The National Oceanic and Atmospheric Administration (NOAA) has released its latest State of the Climate Report. Highlights include:

  • The combined average temperature over global land and ocean surfaces for September 2014 was the highest on record for September, at 0.72°C (1.30°F) above the 20th century average of 15.0°C (59.0°F).

  • The global land surface temperature was 0.89°C (1.60°F) above the 20th century average of 12.0°C (53.6°F), the sixth highest for September on record. For the ocean, the September global sea surface temperature was 0.66°C (1.19°F) above the 20th century average of 16.2°C (61.1°F), the highest on record for September and also the highest on record for any month.

  • The combined global land and ocean average surface temperature for the January–September period (year-to-date) was 0.68°C (1.22°F) above the 20th century average of 14.1°C (57.5°F), tying with 1998 as the warmest such period on record.

Last month, Britain had its driest September since national records began in 1910, with just 20% of the average rainfall for the month. Besides breaking the record itself, this rainfall deficit is especially notable as the preceding eight-month period (January–August) was the wettest such period on record. Meanwhile, 30.6% of the contiguous USA was in drought, with conditions worsening in many regions. Nearly 100% of California and Nevada were in "moderate-to-exceptional" drought.

If 2014 maintains its current trend for the remainder of the year, it will be the warmest calendar year on record, says NOAA. The agency's findings are in strong agreement with both NASA and the JMA, who both reported a record warm September earlier this month too. It also seems quite likely that we'll see an El Niño event during the winter, which could send global temperature anomalies even higher.

 

2014 global warming trend

 

  speech bubble Comments »
 

 

 

21st October 2014

World's first carbon-capture coal power plant

The world’s first commercial-scale carbon capture and storage (CCS) process on a coal-fired power plant has been officially opened at Canada's Boundary Dam Power Station. This $1.4 billion project will cut CO2 emissions from the plant by 90% and sulphur dioxide emissions by 100%.

 

worlds first commercial scale carbon capture coal power plant 2014

 

Electric utility company SaskPower’s new process involves retrofitting an old 110-megawatt (MW) coal-fired plant (that was first commissioned in 1959), adding solvent-based processors to strip away carbon dioxide, and then piping the CO2 to a nearby oil field. When fully optimised, it will capture up to a million tonnes of carbon dioxide annually, the equivalent of taking 250,000 cars off the road. The power unit equipped with CCS technology will continue to use coal to power approximately 100,000 homes and businesses in Saskatchewan, near the Canada-U.S. border. The captured CO2 will be used for enhanced oil recovery, with the remainder stored safely and permanently deep underground and continuously monitored.

The Canadian federal government paid $240 million towards the project. The launch was attended by more than 250 people from over 20 countries representing governments, industries and media. Attendees at the event toured the facility and learned how they can access SaskPower’s expertise and knowledge to develop their own CCS initiatives.

“This project is important because it is applicable to 95% of the world’s coal plants,” said Bill Boyd, Saskatchewan Minister of the Economy. “As nations develop emission regulations, they will come to us to see how we continue to provide affordable coal power to customers, but in an environmentally sustainable way.”

This follows news last month of a similar project being developed in Jacksonville, USA. The Environmental Protection Agency (EPA) approved permits allowing the FutureGen Industrial Alliance to capture and store CO2 deep underground – the first project of its kind in the U.S.

“The opening of this new SaskPower plant reinforces the great innovation and development that can take place if you have strong investment and partnerships from the government and industry,” said U.S. Senator Heidi Heitkamp (D-ND). “From my more than a decade working at Dakota Gasification in North Dakota, and from visiting the construction of the SaskPower facility just over a year ago, I understand just how important it is that we look to the future in how we harness our energy. Coal is a key resource in both Canada and the U.S., and through the development of clean coal technology, we can create North American independence and energy security, while also reducing emissions. We need to develop more clean coal plants to make that possible, and in the U.S., we can learn from the steps Canada has taken to find a realistic path forward for coal.”

The economics of CCS are still a major issue, however. At present, SaskPower's project is expensive and depends on having a nearby source of coal alongside an additional revenue stream from the enhanced oil recovery. Environmentalists have also continued to express concerns.

“At the end of the day, many people are going to wonder why SaskPower is investing $1.4-billion in 'clean coal' technology instead of wind, solar or geothermal energy,” said Victor Lau, Saskatchewan Greens Leader. “Our party will be monitoring future developments of this project very carefully.”

 

  speech bubble Comments »
 

 

 

18th October 2014

Lockheed Martin planning a compact fusion reactor within 10 years

This week, Lockheed Martin announced plans for a small-scale fusion power plant to be developed in as little as 10 years. A number of experts have expressed doubts over its viability.

 

lockheed martin compact fusion reactor design 2019 2024 technology

 

If it ever became a reality, fusion power would be truly world-altering – a clean, safe and essentially limitless supply of energy allowing humanity's continued survival for centuries and millennia to come. The international project known as ITER is planned for operation in 2022 and its eventual successor may emerge in the 2040s. Widespread deployment of fusion is not expected until 2070.

U.S. defence giant Lockheed Martin hopes to accelerate progress in this area, by developing what it calls a compact fusion reactor (CFR). This would be around 10 times smaller than conventional tokamak designs, small enough to fit on the back of a truck and generating 100 megawatts (MW) of power. The company intends to build a prototype within five years – according to its press release – with commercial introduction five years after that. It has several patents pending for the work and is looking for partners in academia, industry and among government laboratories.

As illustrated above, the main improvement over ITER would be the use of a superconducting torus to create a differently shaped magnetic field, able to contain plasma far better than previous configurations. These small reactors could be fitted in U.S. Navy warships and submarines while eliminating the need for other fuel types. They could power small cities of up to 100,000 people, allow planes to fly with unlimited range, or even be used in spacecraft to cut journey times to Mars from six months to a single month. Using a CFR, the cost of desalinated water could fall by 60 percent.

 

 

If this sounds too good to be true, it may well be. Although Lockheed has been successful in its magnetised ion confinement experiments, a number of significant challenges remain for a working prototype with plasma confinement – let alone a commercialised version.

"I think it's very overplayed," University of California nuclear engineering professor Dr. Edward Morse told The Register. "They are being very cagey about divulging details."

"Getting net energy from fusion is such a goddamn difficult undertaking," said University of Texas physicist Dr. Swadesh M. Mahajan, in an interview with Mother Jones. "We know of no materials that would be able to handle anywhere near that amount of heat."

"The nuclear engineering clearly fails to be cost effective," Tom Jarboe told Business Insider in an email.

For these reasons, it is perhaps best to wait for more news and developments before adding the CFR to our timeline. We will, of course, keep you updated on Lockheed's progress as it emerges. You can also discuss this project on our forum.

 

  speech bubble Comments »
 

 

 

14th October 2014

Onshore wind is cheaper than coal, gas and nuclear

Generating electricity from onshore wind is cheaper than gas, coal and nuclear when externalities are stacked with the levelised cost of energy and subsidies, according to a new study ordered and endorsed by the European Commission.

 

onshore wind power

 

A new report by the energy consultancy firm Ecofys has been analysed by the European Wind Energy Association (EWEA). Data in the report shows that onshore wind now has an approximate cost of €105 per megawatt hour (MWh) which is cheaper than gas (up to €164), nuclear (€133) and coal (between €162-233). Offshore wind comes in at €186 and solar PV has a cost of around €217 per MWh.

The total cost of energy production – which factors in externalities such as air quality, climate change and human toxicity among others – shows that coal is more expensive than the highest retail electricity price in the EU. The report puts the figure of external costs of the EU's energy mix in 2012 at between €150 and €310 billion (US$190 and US$394 billion).

Justin Wilkes, deputy chief executive officer of the European Wind Energy Association, said: "This report highlights the true cost of Europe's dependence on fossil fuels. Renewables are regularly denigrated for being too expensive and a drain on the taxpayer. Not only does the Commission's report show the alarming cost of coal but it also presents onshore wind as both cheaper and more environmentally-friendly."

Onshore and offshore wind technologies also have room for significant cost reduction. Coal on the other hand is a fully mature technology and is unlikely to reduce costs any further.

He added: "We are heavily subsidising the dirtiest form of electricity generation while proponents use coal's supposed affordability as a justification for its continued use. The irony is that coal is the most expensive form of energy in the European Union. This report shows that we should use the 2030 climate and energy package as a foundation for increasing the use of wind energy in Europe to improve our competitiveness, security and environment."

 

  speech bubble Comments »
 

 

 

9th October 2014

Fusion reactor concept could be cheaper than coal

The University of Washington is developing a new fusion reactor design that could be one-tenth the cost of ITER – while producing five times the amount of energy.

 

HIT-SI3

 

Fusion energy sounds almost too good to be true – zero greenhouse gas emissions, no long-lived radioactive waste, and a nearly unlimited fuel supply. Perhaps the biggest roadblock to adopting fusion energy is that the economics haven't worked out. Fusion power designs aren't cheap enough to outperform systems that use fossil fuels such as coal and natural gas.

Engineers at the University of Washington (UW) hope to change that. They have designed a concept for a fusion reactor that, when scaled up to the size of a large electrical power plant, would rival costs for a new coal-fired plant with similar electrical output. The team will present its reactor design and cost-analysis findings on 17th October at the Fusion Energy Conference in St. Petersburg, Russia.

“Right now, this design has the greatest potential of producing economical fusion power of any current concept,” says Thomas Jarboe, a UW professor of aeronautics and astronautics and an adjunct professor in physics.

The reactor – called the dynomak – began as a class project taught by Jarboe two years ago. After the class had ended, Jarboe and doctoral student Derek Sutherland (who previously worked on a reactor design at MIT) continued to develop and refine the concept.

The design builds on existing technology and creates a magnetic field within a closed space to hold plasma in place long enough for fusion to occur, allowing the hot plasma to react and burn. The reactor itself would be largely self-sustaining, meaning it would continuously heat the plasma to maintain thermonuclear conditions. Heat generated from the reactor would heat up a coolant that is used to spin a turbine and generate electricity, similar to how a typical power reactor works.

“This is a much more elegant solution, because the medium in which you generate fusion is the medium in which you’re also driving all the current required to confine it,” Sutherland says.

 

fusion design

 

There are several ways to create a magnetic field, which is crucial to keeping a fusion reactor going. The UW’s design is known as a spheromak – meaning it generates the majority of magnetic fields by driving electrical currents into the plasma itself. This reduces the amount of required materials and actually allows researchers to shrink the overall size of the reactor.

Other designs, such as the ITER experimental fusion reactor being built in France – due to be operational in 2022 – have to be much larger than UW’s because they rely on superconducting coils that circle around the outside of the device to provide a similar magnetic field. When compared with the fusion reactor concept in France, the UW’s is much less expensive – about one-tenth the cost of ITER – while producing five times the amount of energy.

The UW researchers factored the cost of building a fusion reactor power plant using their design and compared that with building a coal power plant. They used a metric called “overnight capital costs,” which includes all costs, particularly startup infrastructure fees. A fusion power plant producing a gigawatt (1 billion watts) of power would cost $2.7 billion, while a coal plant of the same output would cost $2.8 billion, according to their analysis.

“If we do invest in this type of fusion, we could be rewarded because the commercial reactor unit already looks economical,” Sutherland said. “It’s very exciting.”

Right now, the UW’s concept is about one-tenth the size and power output of a final product, which is still years away. The researchers have successfully tested the prototype’s ability to sustain plasma efficiently, and as they further develop and expand the size of the device, they can ramp up to higher-temperature plasma and get significant fusion power output. The team has filed patents on the concept with the UW’s Centre for Commercialisation and plans to continue developing and scaling up its prototypes. The research was funded by the U.S. Department of Energy.

 

  speech bubble Comments »
 

 

 

8th October 2014

Ocean warming in Southern Hemisphere has been greatly underestimated

The evidence for global warming continues to pour in. A new study of ocean heat content shows that temperatures have been greatly underestimated in the Southern Hemisphere. As a result, the world's oceans are now absorbing between 24 and 58 per cent more energy than previously thought.

 

ocean heat content southern hemisphere global warming underestimated
Like a fleet of miniature research vessels, more than 3,600 robotic floats provide data on upper layers of the world's ocean currents.

 

Scientists from Lawrence Livermore National Laboratory in California, using satellite observations and a large suite of climate models, have found that long-term ocean warming in the upper 700 metres of Southern Hemisphere oceans has been greatly underestimated.

"This underestimation is a result of poor sampling prior to the last decade, and limitations of the analysis methods that conservatively estimated temperature changes in data-sparse regions," said LLNL oceanographer Paul Durack, lead author of a paper in the 5th October issue of the journal Nature Climate Change.

Ocean heat storage is important because it accounts for over 90 percent of excess heat associated with global warming. The observed ocean and atmosphere warming is a result of continuing greenhouse gas emissions. The Southern Hemisphere oceans make up 60 percent of the world's oceans.

The researchers found that climate models simulating the relative increase in sea surface height between Northern and Southern hemispheres were consistent with highly accurate altimeter observations. However, the simulated upper-ocean warming in Northern and Southern hemispheres was inconsistent with observed estimates of ocean heat content change. These sea level and ocean heat content changes should have been consistent, suggesting that until recent improvements in observational data, Southern Hemisphere ocean heat content changes were underestimated.

Since 2004, automated profiling floats called Argo (pictured above) have been used to measure global ocean temperatures from the surface down to 2,000 m (6,560 ft). These 3,600 floats currently observing the global ocean provide systematic coverage of the Southern Hemisphere for the first time. Argo float data over the last decade, as well as earlier measurements, show that the ocean has been steadily warming, according to Durack.

"The Argo data is really critical," he said. "Estimates that we had until now have been pretty systematically underestimating the changes. Prior to 2004, research has been very limited by poor measurement coverage. Our results suggest that ocean warming has been underestimated by 24 to 58 percent. The conclusion that warming has been underestimated agrees with previous studies. However, it's the first time that scientists have tried to estimate how much heat we've missed."

 

ocean heat content global warming map

 

Given that most of the excess heat associated with global warming is in the oceans, this study has important implications for how scientists view the Earth's overall energy budget. Heat currently stored by the oceans will eventually be released, causing land temperatures to accelerate and triggering more extreme climate events.

"We continue to be stunned at how rapidly the ocean is warming," said Sarah Gille, a Scripps Institution of Oceanography professor who was not involved in the study. "Even if we stopped all greenhouse gas emissions today, we'd still have an ocean that is warmer than the ocean of 1950, and that heat commits us to a warmer climate. Extra heat means extra sea level rise, since warmer water is less dense, so a warmer ocean expands."

"An important result of this paper is the demonstration that the oceans have continued to warm over the past decade, at a rate consistent with estimates of Earth’s net energy imbalance," says Prof. Steve Rintoul, from Australia’s Commonwealth Scientific and Industrial Research Organisation. "While the rate of increase in surface air temperatures slowed in the last 10 to 15 years, the heat stored by the planet, which is heavily dominated by the oceans, has steadily increased as greenhouse gases have continued to rise."

These new results are consistent with another new paper that appears in the same issue of Nature Climate Change. Co-author Felix Landerer of NASA's Jet Propulsion Laboratory, who contributed to both studies, says, "Our other new study on deep-ocean warming found that from 2005 to the present, Argo measurements recorded a continuing warming of the upper-ocean. Using the latest available observations, we're able to show that this upper-ocean warming and satellite measurements are consistent."

In related news, a report by Edinburgh's Heriot-Watt University – based on the work of 30 experts – finds that ocean acidification has increased by 26% since pre-industrial times. It is now causing nearly $1 trillion of damage to coral reefs each year, threatening the livelihoods of 400 million people.

 

  speech bubble Comments »
 

 

 

3rd October 2014

Eastern basin of the Aral Sea has completely dried up

This year marks another milestone for the Aral Sea — a once huge lake in Central Asia that has been shrinking rapidly since the 1960s. For the first time in modern history, its eastern basin has completely dried up.

 

aral sea eastern basin dried up 2000 2014

 

These images, taken by NASA's flagship Terra satellite, show how the Aral Sea has changed in just 14 years. It is now apparent that its eastern basin has completely dried up. The transformation is especially stark when compared to the approximate shoreline location in 1960 (black outline).

"This is the first time the eastern basin has completely dried in modern times," says Philip Micklin, a geographer from Western Michigan University and expert on the Aral Sea. "And it is likely the first time it has completely dried in 600 years, since Medieval desiccation associated with diversion of Amu Darya to the Caspian Sea."

In the 1950s and 60s, the government of the former Soviet Union diverted the Amu Darya and the Syr Darya – the region's two major rivers – in order to irrigate farmland. This diversion began the lake's gradual retreat. By the year 2000, the lake had separated into the North (Small) Aral Sea in Kazakhstan and the South (Large) Aral Sea in Uzbekistan. The South Aral had further split into western and eastern lobes.

 

abandoned boats in the aral sea
The rusting remains of abandoned boats in the Aral Sea, Kazakhstan.

 

The eastern lobe of the South Aral nearly dried in 2009, then saw a huge rebound in 2010. Water levels continued to fluctuate annually in alternately dry and wet years.

According to Micklin, the desiccation in 2014 occurred because there has been less rain and snow in the watershed that starts in the Pamir Mountains; this has greatly reduced water flow on the Amu Darya. In addition, huge amounts of river water continue to be withdrawn for irrigation. The Kok-Aral Dam across the Berg Strait – a channel that connects the northern Aral Sea with the southern part – played some role, but has not been a major factor this year, he said.

Formerly the world's fourth largest lake (pictured below in 1964), the Aral Sea is often described as the worst ecological disaster on the planet. With its eastern half now gone, what remains of the western half is expected to vanish by 2019.

 

aral sea in 1964
Satellite view of the Aral Sea in 1964.

 

  speech bubble Comments »
 

 

 

 
     
       
     
   
« Previous  
   
     
   

 
     
 

Blogs

AI & Robotics Biology & Medicine Business & Politics Computers & the Internet
Energy & the Environment Home & Leisure Military & War Nanotechnology
Physics Society & Demographics Space Transport & Infrastructure

 

 

Archive

2014

 

2013

 

2012

 

2011

 

2010

 

 
 
 
 

 


future timeline twitter future timeline facebook group future timeline youtube channel account videos future timeline rss feed