Why the Battery State-of-charge cannot be measured accurately
Measuring stored energy in an electrochemical device, such as a battery, is complex and state-of-charge (SoC) readings on a fuel gauge provide only a rough estimate. Users often compare battery SoC with the fuel gauge of a vehicle. Calculating fluid in a tank is simple because a liquid is a tangible entity; battery state-of-charge is not. Nor can the energy stored in a battery be quantified because prevailing conditions such as load current and operating temperature influence its release. A battery works best when warm; performance suffers when it is cold. In addition, a battery loses capacity through aging.
Current fuel gauge technologies are fraught with limitations and this came to light when users of the new iPad assumed that a 100 percent charge on the fuel gauge should also relate to a fully charged battery. This is not always so and users complained that the battery was only at 90 percent.
The modern fuel gauge used in iPads, smartphones and laptops read SoC through coulomb counting and voltage comparison. The complexity lies in managing these variables when the battery is in use. Applying a charge or discharge acts like a rubber band, pulling the voltage up or down, making a calculated SoC reading meaningless. In open circuit condition, as is the case when measuring a naked battery, a voltage reference may be used; however temperature and battery age will affect the reading. The open terminal voltage as a SoC reference is only reliable when including these environmental conditions and allowing the battery to rest for a few hours before the measurement.
In the case of the iPad, a 10 percent discrepancy between fuel gauge and true battery SoC is acceptable for consumer products. The accuracy will likely drop further with use, and depending on the effectiveness of a self-learning algorithm, battery aging can add another 20-30 percent to the error. By this time the user has gotten used to the quirks of the device and the oddity is mostly forgotten or accepted. While differences in the runtime cause only a mild inconvenience to a casual user, industrial applications, such as the electric powertrain in an electric vehicle, will need a better system. Improvements are in the work, and these developments may one day also benefit consumer products.
Coulomb counting is the heart of today’s fuel gauge. The theory goes back 250 years when Charles-Augustin de Coulomb first established the “Coulomb Rule.” It works on the principle of measuring in-and-out flowing currents. Coulomb counting also produces errors; the outflowing energy is always less than what goes in. Inefficiencies in charge acceptance, especially towards the end of charge, tracking errors, as well as losses during discharge and self-discharge while in storage contribute to this. Self-learning and periodic calibrations through a full charge/discharge assure an accuracy most can live with.
About the Author
Isidor Buchmann is the founder and CEO of Cadex Electronics Inc. For three decades, Buchmann has studied the behavior of rechargeable batteries in practical, everyday applications, has written award-winning articles including the best-selling book “Batteries in a Portable World,” now in its third edition. Cadex specializes in the design and manufacturing of battery chargers, analyzers and monitoring devices. For more information on batteries, visit www.batteryuniversity.com; product information is on www.cadex.com.
Cadex Electronics Inc., www.BatteryUniversity.com; firstname.lastname@example.org
Posted by Janine E. Mooney, Editor
April 04, 2012