Cognitive computing


Cognitive computing (CC) refers to technology platforms that, broadly speaking, are based on the scientific disciplines of artificial intelligence and signal processing. These platforms encompass machine learning, reasoning, natural language processing, speech recognition and vision (object recognition), human–computer interaction, dialog and narrative generation, among other technologies.[1][2]


At present, there is no widely agreed upon definition for cognitive computing in either academia or industry.[1][3][4]

In general, the term cognitive computing has been used to refer to new hardware and/or software that mimics the functioning of the human brain[5][6][7][8][9] (2004) and helps to improve human decision-making.[10] In this sense, CC is a new type of computing with the goal of more accurate models of how the human brain/mind senses, reasons, and responds to stimulus. CC applications link data analysis and adaptive page displays (AUI) to adjust content for a particular type of audience. As such, CC hardware and applications strive to be more affective and more influential by design.

Basic scheme of a cognitive system. With sensors, such as keyboards, touchscreens, cameras, microphones or temperature sensors, signals from the real world environment can be detected. For perception, these signals are recognised by the cognition of the cognitive system and converted into digital information. This information can be documented and is processed. The result of deliberation can also be documented and is used to control and execute an action in the real world environment with the help of actuators, such as engines, loudspeakers, displays or air conditioners for example.

Cognitivie systems may be engineered to feed on dynamic data in real-time, or near real-time,[11] and may draw on multiple sources of information, including both structured and unstructured digital information, as well as sensory inputs (visual, gestural, auditory, or sensor-provided).[12]

Cognitive analyticsEdit

Cognitive computing-branded technology platforms typically specialize in the processing and analysis of large, unstructured datasets.[13]


Even if cognitive computing can not take the place of teachers, it can still be a heavy driving force in the education of students. Cognitive computing being used in the classroom is applied by essentially having an assistant that is personalized for each individual student. This cognitive assistant can relieve the stress that teachers face while teaching students, while also enhancing the student’s learning experience over all.[14] Teachers may not be able to pay each and every student individual attention, this being the place that cognitive computers fill the gap. Some students may need a little more help with a particular subject. For many students, Human interaction between student and teacher can cause anxiety and can be uncomfortable. With the help of Cognitive Computer tutors, students will not have to face their uneasiness and can gain the confidence to learn and do well in the classroom.[15] While a student is in class with their personalized assistant, this assistant can develop various techniques, like creating lesson plans, to tailor and aid the student and their needs.
Numerous tech companies are in the process of developing technology that involves cognitive computing that can be used in the medical field. The ability to classify and identify is one of the main goals of these cognitive devices.[16] This trait can be very helpful in the study of identifying carcinogens. This cognitive system that can detect would be able to assist the examiner in interpreting countless numbers of documents in a lesser amount of time than if they did not use Cognitive Computer technology. This technology can also evaluate information about the patient, looking through every medical record in depth, searching for indications that can be the source of their problems.
Together with Artificial Intelligence, it has been used in warehouse management systems  to collect, store, organize and analyze all related supplier data. All these aims at improving efficiency, enabling faster decision-making, monitoring inventory and fraud detection[17]
Other use cases

Industry workEdit

Cognitive computing in conjunction with big data and algorithms that comprehend customer needs, can be a major advantage in economic decision making.

The powers of cognitive computing and artificial intelligence hold the potential to affect almost every task that humans are capable of performing. This can negatively affect employment for humans, as there would be no such need for human labor anymore. It would also increase the inequality of wealth; the people at the head of the cognitive computing industry would grow significantly richer, while workers without ongoing, reliable employment would become less well off.[18]

The more industries start to utilize cognitive computing, the more difficult it will be for humans to compete.[18] Increased use of the technology will also increase the amount of work that AI-driven robots and machines can perform. Only extraordinarily talented, capable and motivated humans would be able to keep up with the machines. The influence of competitive individuals in conjunction with AI/CC with has the potential to change the course of humankind.[19]

See alsoEdit


  1. ^ a b Kelly III, Dr. John (2015). "Computing, cognition and the future of knowing" (PDF). IBM Research: Cognitive Computing. IBM Corporation. Retrieved February 9, 2016.
  2. ^ Augmented intelligence, helping humans make smarter decisions. Hewlett Packard Enterprise. Archived April 27, 2016, at the Wayback Machine
  3. ^ "Cognitive Computing". April 27, 2014.
  4. ^ Gutierrez-Garcia, J. Octavio; López-Neri, Emmanuel (November 30, 2015). "Cognitive Computing: A Brief Survey and Open Research Challenges". 2015 3rd International Conference on Applied Computing and Information Technology/2nd International Conference on Computational Science and Intelligence: 328–333. doi:10.1109/ACIT-CSI.2015.64. ISBN 978-1-4673-9642-4. S2CID 15229045.
  5. ^ Terdiman, Daniel (2014) .IBM's TrueNorth processor mimics the human brain.
  6. ^ Knight, Shawn (2011). IBM unveils cognitive computing chips that mimic human brain TechSpot: August 18, 2011, 12:00 PM
  7. ^ Hamill, Jasper (2013). Cognitive computing: IBM unveils software for its brain-like SyNAPSE chips The Register: August 8, 2013
  8. ^ Denning. P.J. (2014). "Surfing Toward the Future". Communications of the ACM. 57 (3): 26–29. doi:10.1145/2566967. S2CID 20681733.
  9. ^ Dr. Lars Ludwig (2013). "Extended Artificial Memory. Toward an integral cognitive theory of memory and technology" (pdf). Technical University of Kaiserslautern. Retrieved February 7, 2017. {{cite journal}}: Cite journal requires |journal= (help)
  10. ^ "Automate Complex Workflows Using Tactical Cognitive Computing: Coseer". Retrieved July 31, 2017.
  11. ^ Ferrucci, David; Brown, Eric; Chu-Carroll, Jennifer; Fan, James; Gondek, David; Kalyanpur, Aditya A.; Lally, Adam; Murdock, J. William; Nyberg, Eric; Prager, John; Schlaefer, Nico; Welty, Chris (July 28, 2010). "Building Watson: An Overview of the DeepQA Project" (PDF). AI Magazine. 31 (3): 59–79. doi:10.1609/aimag.v31i3.2303. S2CID 1831060. Archived from the original (PDF) on February 28, 2020.
  12. ^ Deanfelis, Stephen (2014). Will 2014 Be the Year You Fall in Love With cognitive computing? Wired: 2014-04-21
  13. ^ "Cognitive analytics - The three-minute guide" (PDF). 2014. Retrieved August 18, 2017.
  14. ^ Sears, Alec (April 14, 2018). "The Role Of Artificial Intelligence In The Classroom". ElearningIndustry. Retrieved April 11, 2019.
  15. ^ Coccoli, Mauro; Maresca, Paolo; Stanganelli, Lidia (May 21, 2016). "Cognitive computing in education". Journal of e-Learning and Knowledge Society. 12 (2).
  16. ^ Dobrescu, Edith Mihaela; Dobrescu, Emilian M. (2018). "Artificial Intelligence (Ai) - The Technology That Shapes The World" (PDF). Global Economic Observer. 6 (2): 71–81. ProQuest 2176184267.
  17. ^ "Smart Procurement Technologies for the Construction Sector". Retrieved March 2, 2022.
  18. ^ a b Makridakis, Spyros (June 2017). "The forthcoming Artificial Intelligence (AI) revolution: Its impact on society and firms". Futures. 90: 46–60. doi:10.1016/j.futures.2017.03.006.
  19. ^ West, Darrell M. (2018). The Future of Work: Robots, AI, and Automation. Brookings Institution Press. ISBN 978-0-8157-3293-8. JSTOR 10.7864/j.ctt1vjqp2g.[page needed]

Further readingEdit

  • Russell, John (February 15, 2016). "Mapping Out a New Role for Cognitive Computing in Science". HPCwire. Retrieved April 21, 2016.