It looks like you're using an Ad Blocker.
Please white-list or disable AboveTopSecret.com in your ad-blocking tool.
Thank you.
Some features of ATS will be disabled while you continue to use an ad-blocker.
originally posted by: neoholographic
This makes no sense. You have to structure the problem for humans when they play games and most of the times humans have instructions on how to play the game.
When AlphaGo won in the game of Go it was HUGE. In this area, it's obvious you don't understand what took place. The reason Elon Musk and and the C.E.O. of DeepMind were so excited about this is because they thought these milestones were 5 to 10 years away.
The fact that you try to act like this is something that's just so simple shows your ignorance in this area. If it was so simple, why aren't you creating an A.I. company and selling it for $500 million?
The reason AlphaGo was seen as such a milestone is because it did something very important. It made itself better without human intervention.
originally posted by: neoholographic
Yet, just about every researcher in the field of A.I. said it was a big breakthrough and it's something they didn't think would happen until 5 to 10 years down the road.
Why should anyone believe you over the C.E.O. of DeepMind and other A.I. Researchers?
You have shown in this thread you don't understand the issue and you don't even understand how A.I. and Big Data are connected.
You keep saying it's not intelligent but you haven't even defined what you mean by intelligence. You keep making these general statements that haven't refuted anything that has been said in this thread.
a (1) : the ability to learn or understand or to deal with new or trying situations : reason; also : the skilled use of reason (2) : the ability to apply knowledge to manipulate one's environment or to think abstractly as measured by objective criteria (as tests)
AI learns to solve quantum state of many particles at once
The same type of artificial intelligence that mastered the ancient game of Go could help wrestle with the amazing complexity of quantum systems containing billions of particles.
Google’s AlphaGo artificial neural network made headlines last year when it bested a world champion at Go. After marvelling at this feat, Giuseppe Carleo of ETH Zurich in Switzerland thought it might be possible to build a similar machine-learning tool to crack one of the knottiest problems in quantum physics.
Now, he has built just such a neural network – which could turn out to be a game changer in understanding quantum systems.
Go is far more complex than chess, in that the number of possible positions on a Go board could exceed the number of atoms in the universe. That’s why an approach based on brute-force calculation, while effective for chess, just doesn’t work for Go.
In that sense, Go resembles a classic problem in quantum physics: how to describe a quantum system that consists of many billions of atoms, all of which interact with each other according to complicated equations.
“It’s like having a machine learning how to crack quantum mechanics, all by itself,” Carleo says. “I like saying that we have a machine dreaming of Schrödinger’s cat.”
The test contains three categories of questions: logic questions (patterns in sequences of images); mathematical questions (patterns in sequences of numbers); and verbal reasoning questions (questions dealing with analogies, classifications, synonyms and antonyms). Computers have never been too successful at solving problems belonging to the final category, verbal reasoning, but the machine built for this study actually outperformed the average human on these questions.
The researchers had the deep learning machine and 200 human subjects at Amazon’s Mechanical Turk crowdsourcing facility answer the same verbal questions. The result: their system performed better than the average human.
Artificial intelligence is intelligence in machines. It is commonly implemented in computer systems using program software.
Artificial intelligence (or AI) is both the intelligence of machines and the branch of computer science which aims to create it, through "the study and design of intelligent agents"[30] or "rational agents", where an intelligent agent is a system that perceives its environment and takes actions which maximize its chances of success.[31] Achievements in artificial intelligence include constrained and well-defined problems such as games, crossword-solving and optical character recognition and a few more general problems such as autonomous cars.[32] General intelligence or strong AI has not yet been achieved and is a long-term goal of AI research.
Go is far more complex than chess, in that the number of possible positions on a Go board could exceed the number of atoms in the universe. That’s why an approach based on brute-force calculation, while effective for chess, just doesn’t work for Go.
You have A.I. systems that learn how to play games without instructions.
...
You have A.I. systems that take I.Q. tests
Progress in artificial intelligence causes some people to worry that software will take jobs such as driving trucks away from humans. Now leading researchers are finding that they can make software that can learn to do one of the trickiest parts of their own jobs—the task of designing machine-learning software.
In one experiment, researchers at the Google Brain artificial intelligence research group had software design a machine-learning system to take a test used to benchmark software that processes language. What it came up with surpassed previously published results from software designed by humans.
In recent months several other groups have also reported progress on getting learning software to make learning software. They include researchers at the nonprofit research institute OpenAI (which was cofounded by Elon Musk), MIT, the University of California, Berkeley, and Google’s other artificial intelligence research group, DeepMind.
originally posted by: neoholographic
Yes it was and this is why people in the field of Physics are so excited. Again, these are real world problems that will be solved by a system like AlphaGo. It's ASININE to try to act like this is just something simple. It's easy to say this on a vacuum on a message board but again, you're not providing a shred of evidence to refute what has been said.
This is a huge steps towards strong artificial intelligence.
You have Artificial Intelligence that outperforms humans on an I.Q. tests.
originally posted by: neoholographic
Big Data isn't a buzzword, it describes a reality that the world of science is facing with the real growth of information.
This is because the algorithm made newer and better versions of itself. The whole idea behind A.I. and the intelligence explosion is that intelligent algorithms will quickly become super intelligent when it's able to create better versions of itself. This is exactly what happened with AlphaGo.
Otkrist Gupta, a researcher at the MIT Media Lab, believes that will change. He and MIT colleagues plan to open-source the software behind their own experiments, in which learning software designed deep-learning systems that matched human-crafted ones on standard tests for object recognition.
- www.technologyreview.com...
During the entire training process (starting at = 1.0), we maintain a replay dictionary which stores (i) the network topology and (ii) prediction performance on a validation set, for all of the sampled models. If a model that has already been trained is re-sampled, it is not re-trained, but instead the previously found validation accuracy is presented to the agent. After each model is sampled an trained, the agent randomly samples 100 models from the replay dictionary and applies the Q-value update defined in Equation 3 for all transitions in each sampled sequence. The Q-value update is applied to the transitions in temporally reversed order, which has been shown to speed up Q-value convergence (Lin, 1993).
During the model exploration phase, we trained each network topology with a quick and aggressive training scheme. For each experiment, we created a validation set by randomly taking 5,000 samples from the training set such that the resulting class distributions were unchanged. For every network, a dropout layer was added after every two layers. The i-th dropout layer, out of a total n dropout layers, had a dropout probability of i/2n. Each model was trained for a total of 20 epochs with the Adam optimizer (Kingma & Ba, 2014) with β1 = 0.9, β2 = 0.999, ε = 10^−8. The batch size was set to 128, and the initial learning rate was set to 0.001. If the model failed to perform better than a random predictor after the first epoch, we reduced the learning rate by a factor of 0.4 and restarted training, for a maximum of 5 restarts. For models that started learning (i.e., performed better than a random predictor), we reduced the learning rate by a factor of 0.2 every 5 epochs. All weights were initialized with Xavier initialization (Glorot & Bengio, 2010). Our experiments using Caffe (Jia et al., 2014) took 8-10 days to complete for each dataset with a hardware setup consisting of 10 NVIDIA GPUs.
After the agent completed the schedule (Table 2), we selected the top ten models that were found over the course of exploration. These models were then finetuned using a much longer training schedule, and only the top five were used for ensembling. We now provide details of the datasets
and the finetuning process.
One set of experiments from Google’s DeepMind group suggests that what researchers are terming “learning to learn” could also help lessen the problem of machine-learning software needing to consume vast amounts of data on a specific task in order to perform it well.
- www.technologyreview.com...
The key result, which emerges naturally from the setup rather than being specially engineered, is that the recurrent network dynamics learn to implement a second RL procedure, independent from and potentially very different from the algorithm used to train the network weights. Critically, this learned RL algorithm is tuned to the shared structure of the training tasks. In this sense, the learned algorithm builds in domain-appropriate biases, which can allow it to operate with greater efficiency than a general-purpose algorithm.
originally posted by: Protector
So the AI determines a more optimal Reinforcement Learning (RL) approach than its own initial general conditions--that is, tweaks its own variables to bias itself away from its initial setup to optimize for a specific domain. That could be an advantage in an ideal scenario. Although, it also sounds like it'd have greater problems with any input data (training set) outside of its optimized domain. Meaning, the general-purpose algorithm would probably be more optimal in those cases.
In recent years deep reinforcement learning (RL) systems have attained superhuman performance in a number of challenging task domains. However, a major limitation of such applications is their demand for massive amounts of training data. A critical present objective is thus to develop deep RL methods that can adapt rapidly to new tasks. In the present work we introduce a novel approach to this challenge, which we refer to as deep meta-reinforcement learning. Previous work has shown that recurrent networks can support meta-learning in a fully supervised context. We extend this approach to the RL setting. What emerges is a system that is trained using one RL algorithm, but whose recurrent dynamics implement a second, quite separate RL procedure. This second, learned RL algorithm can differ from the original one in arbitrary ways. Importantly, because it is learned, it is configured to exploit structure in the training domain. We unpack these points in a series of seven proof-of-concept experiments, each of which examines a key aspect of deep meta-RL. We consider prospects for extending and scaling up the approach, and also point out some potentially important implications for neuroscience.
Bengio says the more potent computing power now available, and the advent of a technique called deep learning, which has sparked recent excitement about AI, are what’s making the approach work. But he notes that so far it requires such extreme computing power that it’s not yet practical to think about lightening the load, or partially replacing, machine-learning experts.
Google Brain’s researchers describe using 800 high-powered graphics processors to power software that came up with designs for image recognition systems that rivaled the best designed by humans.
CONCLUSION
A current challenge in artificial intelligence is to design agents that can adapt rapidly to new tasks by leveraging knowledge acquired through previous experience with related activities. In the present work we have reported initial explorations of what we believe is one promising avenue toward this goal. Deep meta-RL involves a combination of three ingredients: (1) Use of a deep RL algorithm to train a recurrent neural network, (2) a training set that includes a series of interrelated tasks, (3) network input that includes the action selected and reward received in the previous time interval. The key result, which emerges naturally from the setup rather than being specially engineered, is that the recurrent network dynamics learn to implement a second RL procedure, independent from and potentially very different from the algorithm used to train the network weights. Critically, this learned RL algorithm is tuned to the shared structure of the training tasks. In this sense, the learned algorithm builds in domain-appropriate biases, which can allow it to operate with greater efficiency than a general-purpose algorithm.
This bias effect was particularly evident in the results of our experiments involving dependent bandits (sections 3.1.2 and 3.1.3), where the system learned to take advantage of the task’s covariance structure; and in our study of Harlow’s animal learning task (section 3.2.2), where the recurrent network learned to exploit the task’s structure in order to display one-shot learning with complex novel stimuli.
Humanity has finally folded under the relentless pressure of an artificial intelligence named Libratus in a historic poker tournament loss. As poker pro Jason Les played his last hand and leaned back from the computer screen, he ventured a half-hearted joke about the anticlimactic ending and the lack of sparklers. Then he paused in a moment of reflection.
“120,000 hands of that,” Les said. “Jesus.”
Even more important, the victory demonstrates how AI has likely surpassed the best humans at doing strategic reasoning in “imperfect information” games such as poker. The no-limit Texas Hold’em version of poker is a good example of an imperfect information game because players must deal with the uncertainty of two hidden cards and unrestricted bet sizes. An AI that performs well at no-limit Texas Hold’em could also potentially tackle real-world problems with similar levels of uncertainty.
“The algorithms we used are not poker specific,” Sandholm explains. “They take as input the rules of the game and output strategy.”
In other words, the Libratus algorithms can take the “rules” of any imperfect-information game or scenario and then come up with its own strategy. For example, the Carnegie Mellon team hopes its AI could design drugs to counter viruses that evolve resistance to certain treatments, or perform automated business negotiations. It could also power applications in cybersecurity, military robotic systems, or finance.