How do we define "Big Data"?

How do we define "Big Data"?

Every day there seems to be another article spiced up with superlatives about Big Data. After all, it's predicted that by 2017 the U.S. could face a shortage of almost 200,000 people with “deep analytical skills.” While the number of headlines about machine learning might lead one to think that we just discovered something profoundly new, the reality is that the technology is nearly as old as computing.

As early as the rise of the Greek city-states of Athens and Sparta to the modern era, we've dreamt about "What if machines could think?" From our science fiction to our research labs, we have long questioned whether the creation of artificial versions of ourselves will somehow help us uncover the origin of our own consciousness, and more broadly, our role on earth. Unfortunately, the learning curve on AI is really damn steep.

Take Turing's 1950s "Turing Test" and how it's popularized the idea of making machines that can think. Back in the 50s, our computing power was limited, we didn’t have access to big-data, and our algorithms were rudimentary. This meant that our ability to advance machine learning research was quite limited. However, that didn’t stop people from trying. In fact, the beauty of machine learning is that instead of pretending computers are human and simply feeding them with knowledge, we help computers to reason and then let them generalize what they’ve learned to new information.

While not well understood, neural networks, deep learning, and reinforcement learning are all machine learning. They’re all methods of creating generalized systems that can perform analysis on new data.

Put a different way, machine learning is one of many artificial intelligence techniques, and things like neural networks and deep learning are just tools that can be used to build better frameworks with broader applications. Back in 1952, Arthur Samuel made a chess program using a very basic form of AI called "alpha beta pruning". This is a method for reducing computational load when working with search trees that represent data, but it’s not always the best strategy for every problem. Even neural networks showed their face in yesteryear with Frank Rosenblatt’s perceptron.

A Complex Model

The perceptron was way ahead of its time, leveraging neuroscience to advance machine learning.

To understand what it’s doing, you first have to understand that most machine learning problems can be broken down into either classification or regression. Classifiers are used to categorize data, while regression models broadly deal with extrapolating out trends to make predictions.

The perceptron is an example of a classifier; it takes a set of data and splits it into multiple sets. In this case, the existence of two traits with respective weights is enough for this object to be classified in the “green” category. Classifiers today separate the spam from your inbox and detect fraud for your bank. (Nifty PDF here.)

This model uses a series of inputs, think features like length, weight, color, and assigns each of them a weight. The model then continuously adjusts the weights until an output is reached that falls within an accepted margin of error.

For example, one could input that the weight of an object that happens to be an apple is 100 grams. The computer doesn’t know it’s an apple, but the perceptron can classify the object as an apple-like-object or a non-apple-like-object by adjusting the classifier’s weights with respect to a known training set of data. Once the classifier has been tuned, it can ideally be reused on a data set it has never been exposed to before to classify unknown objects.

Confused? Let me simplify it.

The perceptron is just one example of many early advances made in machine learning. Neural networks are sort of like big collections of perceptrons working together, a lot like how our brains and neurons work, which is where the name comes from.

Skipping forward a few decades, advancements in AI have continued to be about replicating the way the mind works rather than simply replicating what we perceive its contents to be. Basic, or “shallow”, neural networks are still in use today, but deep learning has caught on as the next big thing. Deep learning models are neural networks with more layers. A totally reasonable reaction to this incredibly unsatisfying explanation is to ask what I mean by layers.

To understand this, we have to remember that just because we say a computer can organize cats and humans into two different groups, the computer itself doesn’t process the task the same way a human would. Machine learning frameworks take advantage of the idea of abstraction to accomplish tasks.

To a human, faces have eyes. To a computer, faces have pixels that are light and dark that make up some abstraction of lines. Each layer of a deep learning model lets the computer identify another level of abstraction of the same object.

This Means "Game On"

So should I get my jacket for the AI winter? Nah.... Despite progress, scientists and entrepreneurs alike have been quick to over-promise the capabilities of AI. The resulting boom and bust cycles are commonly referred to as AI winters.

We have been able to do some unbelievable things with machine learning, like classify objects in video footage for autonomous cars and predict crop yields with satellite imagery. Long short-term memory is helping our machines deal with time-series for things like sentiment analysis in videos. Reinforcement learning, takes ideas from game theory, and includes a mechanism to assist learning through rewards.

That said, despite all progress, the great secret of machine learning is that while we usually know the inputs and outputs of a given problem, and the explicitly programmed code to act as the intermediary, we can’t always identify how the model is going from input to output. Researchers refer to this challenge as the black box problem of machine learning.

Before getting too discouraged, we must remember that the human brain itself is a black box. We don’t really know how it works and cannot examine it at all levels of abstraction. I would be labeled crazy if I asked you to dissect a brain and point to the memories held within it. However, not being able to understand something isn’t game over, it’s game on.

As a member of Relus' recruiting team, Brian Fink focuses on driving talent towards opportunity. Eager to help stretch the professional capabilities of everyone he works with, he's helping startups grow and successfully scale their IT, Recruiting, Big Data, Product, and Executive Leadership teams. An active keynote speaker and commentator, Fink thrives on discovery and building a better-recruiting mousetrap. Follow him on Twitter.





要查看或添加评论,请登录

Brian Fink的更多文章

  • The Future of Technical Recruiting: Adapt or Die Tryin'

    The Future of Technical Recruiting: Adapt or Die Tryin'

    Let’s talk about technical recruiting — the lifeblood of the tech economy, the beating heart of engineering growth, and…

    33 条评论
  • The Only Hiring Question That Matters

    The Only Hiring Question That Matters

    Hiring managers love to complicate things. They’ll wax poetic about “culture fit,” fetishize years of experience like…

    19 条评论
  • Future Vibes in Technical Recruiting

    Future Vibes in Technical Recruiting

    The future of technical recruiting is being rewritten in real-time, and its new protagonist isn’t some Ivy League…

    9 条评论
  • The AI Tidal Wave is Coming for Junior Recruiters

    The AI Tidal Wave is Coming for Junior Recruiters

    Almost a year ago, I made several comments about the future of agency recruiting. Builders like Adam Posner, Matthew…

    48 条评论
  • Stop Being a Bobblehead

    Stop Being a Bobblehead

    HR and Talent Acquisition love to talk about innovation. Every conference, every white paper, every LinkedIn post…

    73 条评论
  • The Art of Engaging Passive Talent: A Data-Driven Guide

    The Art of Engaging Passive Talent: A Data-Driven Guide

    Picture this: A top performer at your competitor—let's call her Emma—is crushing it in her current role. She's not…

    17 条评论
  • Empowerment is The Sauce

    Empowerment is The Sauce

    In a world where most companies are stuck in the past, running on bureaucratic fumes and micromanagement mania, a few…

    13 条评论
  • Agentic AI: Bad Recruiting At Scale?

    Agentic AI: Bad Recruiting At Scale?

    Recruiting is a $200 billion industry that many believe is built on inefficiency, bias, and the illusion of choice…

    55 条评论
  • The “White-Collar Recession” Hiding in a Red-Hot Job Market

    The “White-Collar Recession” Hiding in a Red-Hot Job Market

    Yesterday, you might have caught the Purple Acorn crew talking about some of the recent puzzles in the "Red Hot Job…

    57 条评论
  • Writing: The Refuge of the Rational

    Writing: The Refuge of the Rational

    If life is a warzone, then writing is the foxhole. It’s the sanctuary where thoughts, ideas, and the occasional…

    2 条评论

社区洞察

其他会员也浏览了