Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletter Hub
Free Learning
Arrow right icon
timer SALE ENDS IN
0 Days
:
00 Hours
:
00 Minutes
:
00 Seconds

Tech News - Data

1208 Articles
article-image-amazon-is-the-next-target-on-eu-antitrust-hitlist
Sugandha Lahoti
20 Sep 2018
2 min read
Save for later

Amazon is the next target on EU's antitrust hitlist

Sugandha Lahoti
20 Sep 2018
2 min read
EU Competition Commissioner Margrethe Vestager confirmed that they are doing a preliminary antitrust investigation into Amazon’s business practices on Wednesday. This development was revealed during a press conference hosted to discuss the decision taken on Luxemburg McDonald’s state aid case. Vestager clarified that it is not yet a formal investigation, but the committee is asking questions about how Amazon is using its data. Nevertheless, they have begun probing around, seeking answers to whether Amazon data, collected for legitimate purposes, is also used to give Amazon a competitive advantage over the smaller merchants. The issue, says Vestager, is whether Amazon is using data from the merchants it hosts on its site to secure an advantage in selling products against those same retailers. The regulators want to know whether that data could give Amazon an edge over competitors by providing insight into consumer behavior. “Well, do you then also use this data to do your own calculations? As to what is the new big thing? What is it that people want? What kind of offers do they like to receive? What makes them buy things? And that has made us start a preliminary… antitrust investigation into Amazon’s business practices.” She added that EU regulators have started gathering information on the issue and have sent “quite a number of questionnaires” to merchants and others in order to understand the issue better. Amazon has all the right to fear this investigation. Vestager has the power to fine companies up to 10 percent of their global turnover for breaching EU antitrust rules. Earlier this year, in July, EU had slapped Google with a $5 billion fine for the Android anticompetitive practices. Investors and insiders have been looking for ways to break the company up for a long time. President Donald Trump had also hinted at antitrust action against Amazon, in July. Amazon is yet to comment on this development. What the EU Copyright Directive means for developers – and what you can do Amazon calls Senator Sanders’ claims about ‘poor working conditions’ as “inaccurate and misleading” Amazon hits $1 trillion market value milestone yesterday, joining Apple Inc
Read more
  • 0
  • 0
  • 2324

article-image-in-5-years-machines-will-do-half-of-our-job-tasks-of-today-1-in-2-employees-need-reskilling-upskilling-now-world-economic-forum-survey
Bhagyashree R
20 Sep 2018
5 min read
Save for later

In 5 years, machines will do half of our job tasks of today; 1 in 2 employees need reskilling/upskilling now - World Economic Forum survey

Bhagyashree R
20 Sep 2018
5 min read
Earlier this week, World Economic Forum published a report, The Future of Jobs Report 2018, which is based on a survey they conducted to analyze the trends in the job sector in the period 2018-2022. This survey considered 20 economies and 12 industry sectors. The main focus of this survey was to better understand the potential of new technologies, including automation and algorithms, to create new high-quality jobs and improve the existing job quality and productivity of human employees. Key findings from The Future of Jobs survey #1 Technological advances will drive business growth By 2022, we will see four key technologies enabling business growth: high-speed mobile internet, artificial intelligence, widespread adoption of big data analytics, and cloud technology. 85% of the companies are expected to invest in big data analytics. A large share of companies are also interested in adopting internet of things, app- and web-enabled markets, and cloud computing. Machine learning and augmented and virtual reality will also see considerable business investment. Source: World Economic Forum #2 Acceptance of robots varies across sectors The demand for humanoid robots will be limited in this period, as businesses are more gravitating towards robotics technologies that are at or near commercialization. These technologies include stationary robots, non-humanoid land robots and fully automated aerial drones, in addition to machine learning algorithms and artificial intelligence. Majority of companies (37% to 29%) are showing interest in adopting stationary robots. Oil & Gas industry report the same level of demand for stationary, aerial, and underwater robots. Financial Services industry is planning the adoption of humanoid robots in the period up to 2022. #3 Towards equal work distribution between machines and humans Almost 50% of the companies are expecting that by 2022 automation will lead to some reduction in workforce. While 38% of the companies are more likely to shift their workforce to new productivity-enhancing roles. And more than quarter believe that automation will lead to the creation of new roles in their enterprise. The period from 2018-2022 will see a significant shift in the division of work between humans, machines, and algorithms. Currently, across all the 12 industries surveyed, 71% of the task hours are performed by humans, compared to 29% by machines. By 2022, this average will this average is expected to have shifted to 58% task hours performed by humans and 42% by machines. Source: World Economic Forum Read also: 15 millions jobs in Britain at stake with Artificial Intelligence robots set to replace humans at workforce #4 Emergence of new job opportunities By 2020, with technological advancements newly emerging job roles and opportunities are expected to grow from 16% to 27% of the employee base. The job roles that are affected by technological obsolescence are set to decrease from 31% to 21%. The survey also revealed that there will be a decline of 0.98 million jobs and a gain of 1.74 million jobs. The professions that will enjoy increasing demand include Data Analysts and Scientists, Software and Applications Developers, and Ecommerce and Social Media Specialists. As you can already tell, these are the roles that are significantly based on and enhanced by the use of technology. Read also: Highest Paying Data Science Jobs in 2017 Roles that leverage ‘human' skills are also expected to grow, such as Customer Service Workers, Sales and Marketing Professionals, Training and Development, People and Culture, and Organizational Development Specialists as well as Innovation Managers. Source: World Economic Forum #5 Upskilling and reskilling is the need of the hour With so many businesses embracing technological advancements for business growth, around 54% of the employees will require significant reskilling and upskilling. Out of these 35% are expected to require additional training of up to six months, 9% will require reskilling lasting six to 12 months, while 10% will require additional skills training of more than a year. The key skills that are expected to grow by 2022 include analytical thinking and innovation as well as active learning and learning strategies. Along with these skills, there is an increase in demand for technology design and programming. This indicates a growing demand for various forms of technology competency identified by employers surveyed for this report. Read also: A non programmer’s guide to learning Machine learning Employers are also looking for “human” skills in their employees which include creativity, originality and initiative, critical thinking, persuasion and negotiation. Social influence and emotional intelligence leadership will also see an outsized increase in demand. Read also: 96% of developers believe developing soft skills is important Source: World Economic Forum #6 How companies are planning to address skills gaps To address the skill gaps widened by the adoption of new technologies, companies have highlighted three future strategies. They expect to hire wholly new permanent staff already possessing skills relevant to new technologies, seek to automate the work tasks concerned completely, and retrain existing employees. Read also: Stack skills, not degrees: Industry-leading companies, Google, IBM, Apple no longer require degrees Most companies are considering the option of hiring new permanent staff with relevant skills. A quarter of them are undecided to pursue the retraining of existing employees and two-thirds expect their employees to acquire these skills during their transition period. Between one-half and two-thirds are likely to turn to external contractors, temporary staff and freelancers to address their skills gaps. Source: World Economic Forum Read also: Why learn machine learning as a non-techie? The advancements in technology will come with its own pros and cons. Automation and work augmentation in business will result in decreasing the demand of some of the current job roles. At the same time, this will also open up more opportunities for an entirely new range of livelihood options for workers. To be prepared for this shift, with the help of our employers, we need to upskill ourselves with an agile mindset. To know more in detail, check out the report published by World Economic Forum: The Future of Jobs 2018. Survey reveals how artificial intelligence is impacting developers across the tech landscape Why TensorFlow always tops machine learning and artificial intelligence tool surveys What the IEEE 2018 programming languages survey reveals to us
Read more
  • 0
  • 0
  • 3092

article-image-sap-creates-ai-ethics-guidelines-and-forms-an-advisory-panel
Prasad Ramesh
20 Sep 2018
3 min read
Save for later

SAP creates AI ethics guidelines and forms an advisory panel

Prasad Ramesh
20 Sep 2018
3 min read
“The danger of AI is much greater than the danger of nuclear warheads, by a lot”—Elon Musk SAP, a market leader enterprise software, became the first European technology company to create an AI ethics advisory panel when they made an announcement on Tuesday. They have has announced a set of guiding principles and have developed an external artificial intelligence (AI) ethics advisory panel of five board members. What are the guidelines? The guidelines revolve around recognizing AI’s significant impact on people and society. SAP says that they have designed these guidelines to “help the world run better and improve people’s lives”. The seven guidelines as stated on the SAP website are: We are driven by our values. We design for people. We enable business beyond bias. We strive for transparency and integrity in all that we do. We uphold quality and safety standards. We place data protection and privacy at our core. We engage with the wider societal challenges of artificial intelligence Who is in the AI ethics advisory board? The advisory board panel comprises of experts from various fields like academia, politics and industry. The panel is present for ensuring the adoption of the principles and to develop the principles further collaboratively with the ‘SAP AI steering committee’. The AI ethics panel consists of members who are theology professors, chairmen, law and policy professors, IT professors, and scholars, researchers. The members are: Dr. theol. Peter Dabrock, Chair of Systematic Theology (Ethics), University of Erlangen-Nuernberg Dr. Henning Kagermann, Chairman, acatechBoard of Trustees; acatech Senator Susan Liautaud, Lecturer in Public Policy and Law, Stanford. Founder and Managing Director, Susan Liautaud & Associates Limited (SLAL) Dr. Helen Nissenbaum, Professor, Cornell Tech Information Science Nicholas Wright, Consultant, Intelligent Biology. Affiliated Scholar with Pellegrino Center for Clinical Bioethics Georgetown University Medical Center. An Honorary Research Associate in Institute of Cognitive Neuroscience, University College London Together with the guidelines, SAP’s internal committee and the formed external panel, SAP aims to ensure that the AI capabilities in SAP Leonardo Machine Learning are used to maintain ‘integrity and trust’ in all its solutions. Implementation of AI ethics SAP thinks that the guiding principles also contribute to the AI debate in Europe. Markus Noga, senior vice president, Machine Learning, SAP, is appointed to the high level AI expert group by the European Commission. This European AI expert group was created to design an AI strategy and purpose with ethical guidelines relating to fairness, safety, transparency, by early 2019. Luka Mucic, Chief Financial Officer and member of the Executive Board of SAP Se. stated “SAP considers the ethical use of data a core value. We want to create software that enables the intelligent enterprise and actually improves people’s lives. Such principles will serve as the basis to make AI a technology that augments human talent.” For more information visit the SAP website and read their guiding principles for artificial intelligence. SapFix and Sapienz: Facebook’s hybrid AI tools to automatically find and fix software bugs Sex robots, artificial intelligence, and ethics: How desire shapes and is shaped by algorithms What makes functional programming a viable choice for artificial intelligence projects?
Read more
  • 0
  • 0
  • 3027
Visually different images

article-image-matplotlib-3-0-is-here-with-new-cyclic-colormaps-and-convenience-methods
Natasha Mathur
20 Sep 2018
3 min read
Save for later

Matplotlib 3.0 is here with new cyclic colormaps, and convenience methods

Natasha Mathur
20 Sep 2018
3 min read
Matplotlib team announced Matplotlib version 3.0, on Tuesday. Matplotlib 3.0 comes with new features such as two new cyclic colormaps, AnchoredDirectionArrows feature, and other updates and improvements. Matplotlib is a plotting library for the Python programming language as well as for its numerical mathematics extension, NumPy. It offers an object-oriented API for embedding plots into applications using general-purpose GUI toolkits such as Tkinter, wxPython, Qt, or GTK+. Let’s have a look at what’s new in this latest release. Cyclic Colormaps Two new colormaps namely 'twilight' and 'twilight_shifted' have been added to this new release. These two colormaps start and end on the same color. They have two symmetric halves with equal lightness, but diverging color. AnchoredDirectionArrows added to mpl_toolkits A new mpl_toolkit class AnchoredDirectionArrows, has been added in this release. AnchoredDirectionArrows draws a pair of orthogonal arrows which helps indicate directions on a 2D plot. Several optional parameters can alter the layout of these arrows. For instance, the arrow pairs can be rotated and their color can be changed. The labels and the arrows have the same color by default, but the class may also pass arguments for customizing arrow and text layout. Other than that, the location, length, and width of both the arrows can also be adjusted. Improved default backend selection The default backend needs no longer be set as part of the build process. Instead, builtin backends are tried in sequence at run time, until one of the imports. Also, Headless Linux servers cannot select a GUI backend. Scale axis by a fixed order of magnitude With Matplotlib 3.0, you can scale an axis by a fixed order of magnitude by setting the scilimits argument of Axes.ticklabel_format to the same (non-zero) lower and upper limits. With this setting, the order of magnitude gets adjusted depending on the axis values, rather than remaining fixed. minorticks_on()/off() methods added for colorbar A new method colorbar.Colobar.minorticks_on() has been added in this new release that can correctly display the minor ticks on a colorbar. This method doesn't allow the minor ticks to extend into the regions beyond vmin and vmax. A complementary method named colorbar.Colobar.minorticks_off() has also been added for removing the minor ticks on the colorbar. New convenience methods for GridSpec New convenience methods namely gridspec.GridSpec and gridspec.GridSpecFromSubplotSpec have been added in Matplotlib 3.0. Other Changes Colorbar ticks are now automatic. Legend has a title_fontsize kwarg (and rcParam) now. Multipage PDF support has been added for pgf backend. Pie charts are now circular by default in Matplotlib 3.0 :math: directive has been renamed to :mathmpl: For more information, be sure to check out the official Matplotlib release notes. Creating 2D and 3D plots using Matplotlib How to Customize lines and markers in Matplotlib 2.0 Tinkering with ticks in Matplotlib 2.0
Read more
  • 0
  • 0
  • 4612

article-image-now-deep-reinforcement-learning-can-optimize-sql-join-queries-says-uc-berkeley-researchers
Natasha Mathur
19 Sep 2018
6 min read
Save for later

Now Deep reinforcement learning can optimize SQL Join Queries, says UC Berkeley researchers

Natasha Mathur
19 Sep 2018
6 min read
A team of researchers, Sanjay Krishnan, Zongheng Yang, Ken Goldberg, Joseph M. Hellerstein, and Ion Stoica, from RISELab, UC Berkeley, have shown that deep reinforcement learning is successful at optimizing SQL joins, a problem studied for decades in the database community. Background SQL query optimization has been studied in the database community for almost 40 years. Join Ordering problem is central to query optimization. Despite that problem, there is still a multitude of research projects that attempt to better understand join optimizers performance in terms of multi-join queries. This research shows that reinforcement learning (deep RL) provides a new solution to deal with this problem. The traditional dynamic programs generally reuse the results that have been previously computed via memoization ( optimization technique used primarily to speed up computer programs) while Reinforcement Learning represents the information in the previously computed results with the help of a learned model. “We apply an RL model with a particular structure, a regression problem that associates the downstream value (future cumulative cost) with a decision (a join of two relations for a given query). Training this model requires observing join orders and their costs sampled from a workload. By projecting the effects of a decision into the future, the model allows us to dramatically prune the search space without exhaustive enumeration,” reads the research paper. The research was carried out mainly with two methods, namely, join ordering problem as a Markov Decision process and a deep reinforcement learning optimizer, DQ. Join Ordering using Reinforcement Learning Researchers formulated the join ordering problem as a Markov Decision Process (MDP), which formalizes a wide range of problems such as path planning and scheduling. Then a popular RL technique, Q-learning is applied to solve the join-ordering MDP, where: States, G: the remaining relations to be joined. Actions, c: a valid join out of the remaining relations. Next states, G’: naturally, this is the old “remaining relations” set with two relations removed and their resultant join added. Reward, J: estimated cost of the new join. The Q-function can be defined as Q(G, c), which evaluates the long-term cost of each join .i.e. the cumulative cost for all subsequent joins after the current join decision. Q(G, c) = J(c) + \min_{c’} Q(G’, c’). After access to the Q-function, joins can be ordered in a greedy fashion, which involves starting with an initial query graph, finding the join with the lowest Q(G, c), updating the query graph and repeat. Once done with this, Bellman’s Principle of Optimality is used which tells us if an algorithm is provably optimal. Bellman’s “Principle of Optimality” is one of the most important results in computing. Figure 1: Using a neural network to approximate the Q-function. The output layer, intuitively, means “If we make a join c on the current query graph G, how much does it minimize the cost of the long-term join plan?” Now, as there is no access provided to the true Q-function, it is approximated with the help of a neural network (NN). When an NN is used to learn the Q-function, the technique is called Deep Q-network (DQN). Now,  a neural net is trained that takes in (G, c) and outputs an estimated Q(G,c). DQ, the Deep Reinforcement Learning Optimizer The deep RL-based optimizer uses only a moderate amount of training data to achieve plan costs within 2x of the optimal solution on all cost models. DQ uses a multi-layer perceptron (MLP) neural network which is used to represent the Q-function. First, let's start with collecting data to learn the Q function. Here the past execution data is observed. DQ accepts a list of (G, c, G’, J) from any underlying optimizer. The second step is the featurization of states and actions. Now a neural net is used to represent Q(G, c), it is necessary to feed states G and actions c into the network as fixed-length feature vectors. This featurization process is simple where 1-hot vectors are used for encoding the set of all attributes that exist in the query graph. The third step is Neural network training & planning. Here, DQ makes use of a simple 2-layer fully connected network, by default. Training is done using a standard stochastic gradient descent. Once the Neural Network is trained, DQ accepts an SQL query which is in plain text. It then parses the SQL query into an abstract syntax tree form, featurizes the tree, and invokes the neural network when a candidate join is scored. Lastly, DQ can be periodically re-tuned with the feedback from real execution. Evaluating DQ For this, the researchers used a recently published Join Order Benchmark (JOB). The database comprises 21 tables from IMDB with 33 query templates and a total of 113 queries. Sizes of joins in the queries range from 5 to 15 relations. The training data is collected by DQ from exhaustive enumeration for cases when the number of relations to join is no larger than 10 as well as from greedy algorithm for additional relations. DQ is then compared against several heuristic optimizers (QuickPick; KBZ) as well as classical dynamic programs (left-deep; right-deep; zig-zag). Then the plans that are produced by each optimizer gets scored and compared to the optimal plans achieved via exhaustive enumeration. After this, to show that a learning-based optimizer can adapt to different environments, three designed cost models are used. Results Across all the cost models, the researchers found that DQ is competitive with the optimal solution without any prior knowledge of the index structure. Furthermore,  DQ produces good plans at a much faster speed as opposed to classical dynamic programs. Also, in the large-join regime, DQ achieves drastic speedups. The largest joins DQ performs 10,000x better as compared to exhaustive enumeration, over 1,000x faster as compared to zig-zag, and more than 10x faster than left/right-deep enumeration. “We believe this is a profound performance argument for such a learned optimizer: it would have an even more unfair advantage when applied to larger queries or executed on specialized accelerators (e.g., GPUs, TPUs)” says by Zongheng Yang in the RISELab Blog. For more details, check out the official research paper. MIT’s Transparency by Design Network: A high-performance model that uses visual reasoning for machine interpretability Swarm AI that enables swarms of radiologists, outperforms specialists or AI alone in predicting Pneumonia How Facebook data scientists use Bayesian optimization for tuning their online systems
Read more
  • 0
  • 0
  • 4038

article-image-github-introduces-experiments-a-platform-to-share-live-demos-of-their-research-projects
Bhagyashree R
19 Sep 2018
2 min read
Save for later

GitHub introduces ‘Experiments’, a platform to share live demos of their research projects

Bhagyashree R
19 Sep 2018
2 min read
Yesterday, GitHub introduced the Experiments platform for sharing demonstrations of their research projects and the idea behind them. With this platform, it aims to give the end users “insight into their research and inspire them to think audaciously about the future of software development”. Why has GitHub introduced ‘Experiments’? Just like Facebook and Google, GitHub regularly conducts research in machine learning, design, and infrastructure. The resultant products are rigorously evaluated for stability, performance, and security. If these products meet the success criteria for product release, they are then released for end users. Experiments will help GitHub share details about their research as they happen. ‘Semantic Code Search’: The first demo published on Experiments The GitHub researchers also published their first demo of an experiment called Semantic Code Search. This system helps you search code on GitHub using natural language. How does Semantic Code Search work? The following diagram shows how Semantic Code Search works: Source: GitHub Step1: Learning representations of code In this step, a sequence-to-sequence model is trained to summarize code by supplying (code, docstring) pairs. The docstring here is the target variable the model is trying to predict. Step 2: Learning representations of text phrases Along with learning representations of code, the researchers wanted to find a suitable representation for short phrases. To achieve this, they trained a neural language model by leveraging the fast.ai library. Using the concat pooling approach, the representations of phrases were extracted from the trained model by summarizing the hidden states. Step 3: Mapping code representations to the same vector-space as text In this step, the code representations learned from step 1 were mapped to the vector space of text. To accomplish this they fine-tuned the code-encoder. Step 4: Creating a semantic search system The last step is to bringing everything together to create a semantic search mechanism. The vectorized version of all code is stored in a database, and nearest neighbor lookups are performed to a vectorized search query. You can read the official announcement at GitHub’s blog. To read in more detail about Semantic Code Search, check out the researchers’ post and also try it on Experiments. Packt’s GitHub portal hits 2,000 repositories GitHub parts ways with JQuery, adopts Vanilla JS for its frontend Github introduces Project Paper Cuts for developers to fix small workflow problems, iterate on UI/UX, and find other ways to make quick improvements
Read more
  • 0
  • 0
  • 2430
Unlock access to the largest independent learning library in Tech for FREE!
Get unlimited access to 7500+ expert-authored eBooks and video courses covering every tech area you can think of.
Renews at $15.99/month. Cancel anytime
article-image-a-new-model-optimization-toolkit-for-tensorflow-can-make-models-3x-faster
Prasad Ramesh
19 Sep 2018
3 min read
Save for later

A new Model optimization Toolkit for TensorFlow can make models 3x faster

Prasad Ramesh
19 Sep 2018
3 min read
Yesterday, TensorFlow introduced a new model optimization toolkit. It is a suite of techniques that both new and experienced developers can leverage to optimize machine learning models. These optimization techniques are suitable for any TensorFlow model and will be particularly of use to developers running TensorFlow Lite. What is model optimization in TensorFlow? Support is added for post-training quantization to the TensorFlow Lite conversion tool. This can theoretically result in up to four times more compression in the data and up to three times faster execution for relevant machine learning models. On quantizing the models they work on, developers will also gain additional benefits of less power consumption. Enabling post-training quantization This quantization technique is integrated into the TensorFlow Lite conversion tool. Initiating is easy. After building a TensorFlow model, you can simple enable the ‘post_training_quantize’ flag in the TensorFlow Lite conversion tool. If the model is saved and stored in saved_model_dir, the quantized tflite flatbuffer can be generated. converter=tf.contrib.lite.TocoConverter.from_saved_model(saved_model_dir) converter.post_training_quantize=True tflite_quantized_model=converter.convert() open(“quantized_model.tflite”, “wb”).write(tflite_quantized_model) There is an illustrative tutorial that explains how to do this. To use this technique for deployment on platforms currently not supported by TensorFlow Lite, there are plans to incorporate it into general TensorFlow tooling as well. Post-training quantization benefits The benefits of this quantization technique include: Approx Four times reduction in model sizes. 10–50% faster execution in models consisting primarily of convolutional layers. Three times the speed for RNN-based models. Most models will also have lower power consumption due to reduced memory and computation requirements. The following graph shows model size reduction and execution time speed-ups for a few models measured on a Google Pixel 2 phone using a single core. We can see that the optimized models are almost four times smaller. Source: Tensorflow Blog The speed-up and model size reductions do not impact the accuracy much. The models that are already small to begin with, may experience more significant losses. Here’s a comparison: Source: Tensorflow Blog How does it work? Behind the scenes, optimizations are run by reducing the precision of the parameters (the neural network weights). The reduction is done from their training-time 32-bit floating-point representations to much smaller and efficient 8-bit integer representations. These optimizations ensure pairing the less precise operation definitions in the resulting model with kernel implementations that use a mix of fixed and floating-point math. This results into executing the heaviest computations quickly, but with lower precision. However, the most sensitive ones are still computed with high precision. This gives little accuracy losses. To know more about model optimization visit the TensorFlow website. What can we expect from TensorFlow 2.0? Understanding the TensorFlow data model [Tutorial] AMD ROCm GPUs now support TensorFlow v1.8, a major milestone for AMD’s deep learning plans
Read more
  • 0
  • 0
  • 4542

article-image-accentures-annual-technology-report-finds-all-five-major-trends-in-2018-being-data-centric
Sugandha Lahoti
19 Sep 2018
4 min read
Save for later

Accenture’s annual Technology report finds all five major trends in 2018 being data centric

Sugandha Lahoti
19 Sep 2018
4 min read
Every year Accenture releases their annual forecast of the technology trends unfolding in the next three years. This year their Accenture Technology Vision 2018  highlights five emerging trends shaping the way technology is increasing businesses’ impact across society. The report covers people on all rungs of the corporate ladder, from strategy to operations, moving closer to the center of people’s lives. Accenture’s Technology Vision comprises a three-year set of technology trends. Accenture Evolution Chart Here are the 5 major trends identified in the 2018 report. Trend #1 CITIZEN AI: Raising AI to benefit business and society As artificial intelligence grows in its capabilities, businesses must move to “raise” their AIs to act as responsible, productive members of society. For businesses, this means deploying AI is no longer just about training it to perform a given task. It’s about “raising” it to act as a responsible representative of the business, and a contributing member of society. Per a survey conducted by Accenture, “72 percent of executives report that their organizations seek to gain customer trust and confidence by being transparent in their AI-based decisions and actions. This will be a crucial step in the integration of AI into society. We call it Citizen AI.” Trend #2 EXTENDED REALITY: The End of Distance Extended Reality XR refers to the spectrum of experiences that blurs the line between the real world and the simulated world. The technology immerses the user through visual, audio, and potentially olfactory and haptic cues. The two major types of XR are virtual reality and augmented reality. The report ascertains that Virtual and augmented reality technologies are removing the distance to people, information, and experiences, transforming the ways people live and work. Per their Survey, “27 percent of executives state it is very important for their organizations to be a pioneer in XR solutions.” Trend #3 DATA VERACITY: The Importance of Trust By transforming themselves to run on data, businesses have created a new kind of vulnerability: inaccurate, manipulated, and biased data that leads to corrupted business insights with a major impact on society. Businesses can address this new vulnerability by building confidence in three key data-focused tenets: Provenance, or verifying the history of data from its origin throughout its life cycle Context, or considering the circumstances around its use Integrity, or securing and maintaining data. Strong cybersecurity and data science capabilities are prerequisites for building a strong data intelligence practice to ensure data veracity. Trend #4 FRICTIONLESS BUSINESS: Built to partner at scale Businesses depend on technology-based partnerships for growth, but their own legacy systems aren’t designed to support partnerships at scale. To fully become the connected Intelligent Enterprise, companies must first re-architect themselves. Organizations must actively reshape their business from top to bottom to meet the challenges of creating and managing relationships at scale. They must adopt microservices architectures, and use blockchain and smart contracts to build a strong foundation for technology-based partnerships. Accenture Survey shows that “ 60% of executives report that blockchain and smart contracts will be critical to their organizations over the next three years.” Trend #5 INTERNET OF THINKING: Creating Intelligent Distributed Systems Businesses are making big bets on intelligent environments via robotics, AI and immersive experiences. But to bring these intelligent environments to life, they must extend their infrastructures into the dynamic, real-world environments. For this, companies must extend compute beyond the cloud, toward the edge of networks. They must explore custom hardware solutions and hardware accelerators that let systems reduce latency and compute limitations. Cloud processing remains appealing for high-value learning, predictions, AI-model generation, and storage in situations that are not time critical. But for real-time action, processing must happen at the edge of networks, where the event is occurring. Read the complete  Accenture Technology Vision 2018 report to learn about the technology trends in detail along with the research methodology and the survey demographics. 18 striking AI Trends to watch in 2018 – Part 1 18 striking AI Trends to watch in 2018 – Part 2 Skill Up is back: 5 things we want to find out
Read more
  • 0
  • 0
  • 1955

article-image-elon-musk-reveals-big-plans-with-neuralink
Guest Contributor
18 Sep 2018
3 min read
Save for later

Elon Musk reveals big plans with Neuralink

Guest Contributor
18 Sep 2018
3 min read
Be it a tweet about taking the company private or smoking weed on a radio show, Elon Musk has been in news for all the wrong reasons recently and he is in news again but this time for what he is best admired as a modern day visionary. As per reports the Tesla and SpaceX founder is working on a 'superhuman' product that will connect your brain to a computer. We all know Musk along with eight others founded a company called Neuralink two years ago. The company has been developing implantable brain–computer interfaces, better known as BCIs. While in the short-term the company’s aim is to use the technology to treat brain diseases, Musk’s eventual goal is human enhancement, which he believes will make us more intelligent and powerful than even AI.  According to hints he gave a week ago, Neuralink may soon be close to announcing a product unlike anything we have seen: A brain computer interface. Appearing on the Joe Rogan Experience podcast last week, Musk stated that he’ll soon be announcing a new product – Neuralink – which will connect your brain to a computer, thus making you superhuman. When asked about Neuralink, Musk said "I think we’ll have something interesting to announce in a few months that’s better than anyone thinks is possible. Best case scenario, we effectively merge with AI. It will enable anyone who wants to have superhuman cognition. Anyone who wants. How much smarter are you with a phone or computer or without? You’re vastly smarter, actually. You can answer any question pretty much instantly. You can remember flawlessly. Your phone can remember videos [and] pictures perfectly. Your phone is already an extension of you. You’re already a cyborg. Most people don’t realise you’re already a cyborg. It’s just that the data rate, it’s slow, very slow. It’s like a tiny straw of information flow between your biological self and your digital self. We need to make that tiny straw like a giant river, a huge, high-bandwidth interface." If we visualize what Musk said, it feels like a scene straight from a Hollywood movie. However, many of the creations, from a decade ago, that were thought to belong solely in the world of science-fiction, have become a  reality now. Musk argues that through our over-dependence on smartphones, we have already taken the first step towards our cyborg future. Neuralink is an attempt to just accelerate the process by leaps and bounds. That's not all, Elon Musk was also quoted saying on CNBC. "If your biological self dies, you can upload into a new unit. Literally, with our Neuralink technology". Read the full news on CNBC. About Author Sandesh Deshpande is currently working as a System Administrator for Packt Publishing. He is highly interested in Artificial Intelligence and Machine Learning. Tesla is building its own AI hardware for self-driving cars Elon Musk’s tiny submarine is a lesson in how not to solve problems in tech. DeepMind, Elon Musk, and others pledge not to build lethal AI
Read more
  • 0
  • 0
  • 6218

article-image-how-facebook-data-scientists-use-bayesian-optimization-for-tuning-their-online-systems
Prasad Ramesh
18 Sep 2018
4 min read
Save for later

How Facebook data scientists use Bayesian optimization for tuning their online systems

Prasad Ramesh
18 Sep 2018
4 min read
Facebook data scientists had released a paper, Constrained Bayesian Optimization with Noisy Experiments in 2017 where they describe using Bayesian optimization to design rounds of A/B tests based on prior test results. An A/B test is a randomized experiment, used to determine which variant of A and B is more "effective". They are used for improving a product. Facebook has a large array of backend systems serving billions of people every day. They have a large number of internal parameters that must be tuned carefully using live, randomized experiments, also known as A/B tests. Individual experiments may take a week or longer, so there is a challenge to optimize a set of parameters with the least number of experiments. Bayesian optimization Bayesian optimization is a technique used to solve optimization problems where the objective function (the online metric of interest) does not have an analytic expression. It can only be evaluated through some time consuming operations like a randomized experiment. Bayesian optimization allows joint tuning of more parameters with fewer experiments compared to a grid search or manual tuning. It also helps in finding better values. The Gaussian process (GP) is a Bayesian model that works well for Bayesian optimization. GP provides good uncertainty estimates of how an online metric varies with the parameters of interest. It is illustrated as follows: Source: Facebook research blog The work in the paper was motivated by several challenges in using Bayesian optimization for tuning online systems. The challenges are noise, constraints, and batch experimentation. In the paper, the authors describe a Bayesian approach for handling observation noise in which they include the posterior uncertainty induced by noise in EI’s expectation. In the paper, they describe a Bayesian approach for handling observation noise. A posterior uncertainty is induced by noise in EI’s expectation. Instead of computing the expectation of I(x) under the posterior of f(x), it is computed under the joint posterior of f(x) and f(x*). This expectation no longer has a closed form like El but can easily draw samples of values at past observations f(x_1), …, f(x_n) from the GP posterior. The conditional distribution f(x) | f(x_1), …, f(x_n) has closed form. The results The approach described in the paper is used to optimize various systems at Facebook. Two such optimizations are described in the paper. The first is to optimize six parameters from one of Facebook’s ranking systems. The second one was to optimize seven numeric compiler flags for the HipHop Virtual Machine (HHVM). The web servers powering Facebook use the HHVM to serve requests. The end goal of this optimization was to reduce CPU usage on the web servers, with a constraint of keeping the peak memory usage less. This following figure shows the CPU usage of each configuration tested. There is a 100 total, it also shows the probability that each point satisfied the memory constraint: Source: Facebook research blog The first 30 iterations were randomly generated configurations depicted as a green line. After this, the Bayesian optimization was used to identify parameter configurations to be evaluated. It was observed that Bayesian optimization was able to find better configurations that are more likely to satisfy the constraints. The findings are that Bayesian optimization is an effective and robust tool for optimizing via noisy experiments. For full details, visit the Facebook research blog. You can also take a look at the research paper. NIPS 2017 Special: A deep dive into Deep Bayesian and Bayesian Deep Learning with Yee Whye Teh Facebook’s Glow, a machine learning compiler, to be supported by Intel, Qualcomm and others “Deep meta reinforcement learning will be the future of AI where we will be so close to achieving artificial general intelligence (AGI)”, Sudharsan Ravichandiran
Read more
  • 0
  • 0
  • 3193
article-image-ibm-faces-age-discrimination-lawsuit-after-laying-off-thousands-of-older-workers-bloomberg-reports
Melisha Dsouza
18 Sep 2018
2 min read
Save for later

IBM faces age discrimination lawsuit after laying off thousands of older workers, Bloomberg reports

Melisha Dsouza
18 Sep 2018
2 min read
IBM is defending itself from yet another round of accusations on its firing practices. Just months after a ProPublica report said the technology giant had dismissed more than 20,000 workers older than 40 in the last five years, Yesterday, Bloomberg reports that IBM is facing another age discrimination lawsuit. The suit that follows heavily on the ProPublica report, alleges that the company systematically fired tens of thousands of older workers in recent years as part of an effort to "make the face of IBM younger." Shannon Liss-Riordan, a lawyer known for battling tech giants over mistreatment of their employees, filed a class-action lawsuit in federal court in Manhattan on behalf of three former IBM employees in their 50s and 60s. They claimed that the tech giant fired them earlier this year based on their age. The three employees declare in the suit that "Over the last several years, IBM has been in the process of systematically laying off older employees in order to build a younger workforce," The complaint also states that the company "discriminates against older workers” and "does not consider them for open positions." Ed Barbini (a spokesman for IBM) said in an emailed statement to multiple news outlets: “Changes in our workforce are about skills, not age,” “since 2010 there is no difference in the age of our U.S. workforce, but the skills profile of our employees has changed dramatically. That is why we have been and will continue investing heavily in employee skills and retraining — to make all of us successful in this new era of technology.’’ In the last decade, IBM has reportedly fired thousands of people in the U.S., Canada, and other high-wage jurisdictions -to cut costs and retool its workforce. If the judge allows a class action lawsuit to proceed, it could lead to a drawn-out and costly court battle. IBM would potentially have to pay hundreds of millions of dollars to its former employees. To know more about this lawsuit, head over to Seattletimes. The Intercept says IBM developed NYPD surveillance tools that let cops pick targets based on skin color IBM Files Patent for “Managing a Database Management System using a Blockchain Database” Say hello to IBM RXN, a free AI Tool in IBM Cloud for predicting chemical reactions
Read more
  • 0
  • 0
  • 2135

article-image-cntk-v2-6-is-here-with-net-support-and-onnx-update-among-others
Natasha Mathur
18 Sep 2018
3 min read
Save for later

CNTK v2.6 is here with .Net support and ONNX update among others

Natasha Mathur
18 Sep 2018
3 min read
Microsoft released the version 2.6 of their popular deep learning framework, CNTK or Microsoft Cognitive Toolkit, last week. CNTK v2.6 explores features such as an added .NET support, efficient group convolution, improved sequential convolution, more operators, and ONNX feature update among others. Added .Net Support The Cntk.Core.Managed library has now been converted to .Net Standard. It now supports .Net Core and .Net Framework applications on both Windows as well as Linux. .Net developers will now be able to restore CNTK Nuget packages. To restore the CNTK Nuget packages, use the new .Net SDK style project file with package management format set to PackageReference. Efficient group convolution With CNTK v2.6, the group convolution has been updated. The updated implementation uses cuDNN7 and MKL2017 APIs directly instead of having to create a sub-graph for group convolution (slicing and splicing). This leads to improved experience in terms of both performance and model size. Improved Sequential Convolution Sequential Convolution implementation has also been updated in CNTK v2.6. The new implementation creates a separate sequential convolution layer. This layer offers support for broader cases, such as, where stride > 1 for the sequence axis. So, if sequential convolution is performed over a batch of one-channel black-and-white images then these images will have the same fixed height of 640 with the width of variable lengths. The width is then represented by the sequential axis. More Operators More support has been added in CNTK v2.6 for operators such as depth_to_space and space_to_depth, Tan and Atan, ELU, and Convolution. depth_to_space and space_to_depth There are breaking changes in the depth_to_space and space_to_depth operators. These two operators are updated to match the ONNX specification. Tan and Atan Support has been added for trigonometric ops Tan and Atan. ELU Support added for alpha attribute in ELU op. Convolution The auto padding algorithms of Convolution have been updated to produce symmetric padding at best effort on CPU, without influencing the final convolution output values. This leads to an increase in the range of cases which could be covered by MKL API and also improves the performance, E.g. ResNet50. Updated ONNX CNTK's ONNX import/export has been updated to support ONNX 1.2. A major update has been added on how the batch and sequence axes are handled in export and import.  CNTK's ONNX BatchNormalization op export/import has been updated to the latest spec. A model domain has been added to the ONNX model export. Support has also been added for exporting alpha attribute in ELU ONNX op. Change in Default arguments order There is a major updated to the arguments property in CNTK python API. The default behavior has been updated so now it returns the arguments in python order instead of in C++ order. Because of this, it will return arguments in the same order as they are fed into ops. Bug Fixes Improved input validation added for group convolution. Validation added for padding channel axis in convolution. Proper initialization added for ONNX TypeStrToProtoMap. The Min/Max import implementation has been updated to handle variadic inputs. There are even more updates that come with CNTK v2.6. For more information on those, check out the CNTK official release notes. The Deep Learning Framework Showdown: TensorFlow vs CNTK Deep Learning with Microsoft CNTK ONNX 1.3 is here with experimental function concept
Read more
  • 0
  • 0
  • 2102

article-image-cmu-students-propose-a-competitive-reinforcement-learning-approach-based-on-a3c-using-visual-transfer-between-atari-games
Bhagyashree R
18 Sep 2018
4 min read
Save for later

CMU students propose a competitive reinforcement learning approach based on A3C using visual transfer between Atari games

Bhagyashree R
18 Sep 2018
4 min read
Earlier this month, some students of Robotics Institute Carnegie Mellon University published a paper proposing a learning approach called competitive reinforcement learning using visual transfer. This method, with the help of asynchronous advantage actor critic (A3C) architecture, generalizes a target game using an agent trained on a source game in Atari. What is A3C architecture? The A3C architecture is an asynchronous variant of the actor-critic model, in which the actor takes in the current environment state and determines the best action to take from there. It consists of four convolutional layers, an LSTM layer, and two fully connected layers to predict actions and value functions of the states. In this architecture, multiple worker agents are trained in parallel, each with their own copy of the model and environment. Advantage here refers to a metric that is set to judge how good the agents' actions were. What is the aim of competitive reinforcement learning? The learning approach introduced in this paper aims to use a reinforcement agent to generalize between two related but vastly different Atari games like Pong-v0 and Breakout-v0. This is done by learning visual mappers: given a frame from the source game, we should be able to generate the analogous frame in the target game. In both these games, a paddle is controlled to hit a ball to obtain a certain objective. Using this method the six actions of Pong-v0 {No Operation, Fire, Right, Left, Right Fire, Left Fire} are mapped to the four actions of Breakout-v0 as {Fire, Fire, Right, Left, Right, Left} respectively. The rewards are mapped directly from source game to target game without any scaling. The source and target environment they experimented on was OpenAI gym. They found underlying similarities between the source and the target game to represent common knowledge using Unsupervised Image-to-image Translation (UNIT) Generative adversarial networks (GANs). The target game competes with its visual representation obtained after using the UNIT GAN as a visual mapper between the source and target game. How competitive reinforcement learning works? The following diagram depicts how knowledge is transferred from source game to target game by competitively and simultaneously fine-tuning the model using two different visual representations of the target game: Source: arXiv First stage: The baseline A3C network is trained for source game (Pong-v0) in the first stage of the training process. The knowledge learned is then transferred from this model to learning to play target game (Breakout-v0). The efficiency of transfer learning method in terms of training time and data efficiency across parallel actor-learners is measured. Second stage: In this stage of the training process, two representations of the target game are used amongst the workers in parallel. The first representation of transfer process uses the target game frames taken directly from the environment. The second representation of transfer process uses the frames learned from the visual mapper (visual analogies between games). The ratio of number of workers that train directly on frames queried from the target game and frames mapped from the source game is a hyperparameter that is determined through experimentation. Results They concluded that it is possible to generate a visual mapper for semantically similar games with the use of UNIT GANs. Learning from two different representations of the same game and using them simultaneously for transfer learning stabilizes the learning curve. Although the workers using representations of the target game obtained from the visual mappers did not perform well in a stand alone setting, they showed improvements when used for the competitive learning. To read more about this learning approach and its efficiency, check out this research paper published by Akshita Mittel, Purna Sowmya Munukutla, and Himanshi Yadav: Visual Transfer between Atari Games using Competitive Reinforcement Learning. “Deep meta reinforcement learning will be the future of AI where we will be so close to achieving artificial general intelligence (AGI)”, Sudharsan Ravichandiran This self-driving car can drive in its imagination using deep reinforcement learning Dopamine: A Tensorflow-based framework for flexible and reproducible Reinforcement Learning research by Google
Read more
  • 0
  • 0
  • 2258
article-image-nuscenes-the-largest-open-source-dataset-for-self-driving-vehicles-by-scale-and-nutonomy
Bhagyashree R
17 Sep 2018
2 min read
Save for later

nuScenes: The largest open-source dataset for self-driving vehicles by Scale and nuTonomy

Bhagyashree R
17 Sep 2018
2 min read
Scale and nuTonomy, the leading players in the self-driving vehicle ecosystems, open-sourced a research dataset named nuScenes last week. According to the companies, this is the largest open source dataset for self-driving vehicles, which includes data from LIDAR, RADAR, camera, IMU, and GPS. nuTonomy with the help of Scale’s Sensor Fusion Annotation API, compiled more than 1,000 20-second clips and 1.4 million images. nuScenes comprises of 400,000 sweeps of LIDARs and 1.1 million three-dimensional boxes detected with the combination of RGB cameras, RADAR, and LIDAR. The collection of this much data was facilitated by six cameras, one LIDAR, five RADARs, GPS, and an inertial measurement sensor. They chose the driving routes in Singapore and Boston to showcase challenging locations, times, and weather conditions. This open-source dataset reportedly surpasses in terms of size and accuracy of common datasets including public KITTI dataset, Baidu ApolloScape dataset, Udacity self-driving dataset, and even the more recent Berkeley DeepDrive dataset. Making this huge dataset available to the users will facilitate training and testing different algorithms for autonomous driving, accurately and quickly. Scale CEO Alexandr Wang said: “We’re proud to provide the annotations … as the most robust open source multi-sensor self-driving dataset ever released. We believe this will be an invaluable resource for researchers developing autonomous vehicle systems, and one that will help to shape and accelerate their production for years to come.” You can read more about nuScenes in this full coverage. To know more about nuScenes check out its website and also see the official announcement by Scale on its Twitter page. Google launches a Dataset Search Engine for finding Datasets on the Internet Ethereum Blockchain dataset now available in BigQuery for smart contract analytics 25 Datasets for Deep Learning in IoT
Read more
  • 0
  • 0
  • 3726

article-image-what-can-we-expect-from-tensorflow-2-0
Savia Lobo
17 Sep 2018
3 min read
Save for later

What can we expect from TensorFlow 2.0?

Savia Lobo
17 Sep 2018
3 min read
Last month, Google announced that the TensorFlow community plans to release a preview of TensorFlow 2.0, later this year. However, the date for the preview release has not been disclosed yet. The 2.0 version will include major highlights such as improved eager execution, improved compatibility, support for more platforms and languages, and much more. Key highlights in Tensorflow 2.0 Eager execution would be an important feature of TensorFlow 2.0. It aids in aligning users’ expectations about the programming model better, with TensorFlow practice. This will thus make TensorFlow easier to learn and apply. This version includes a support for more platforms and languages. It will provide an improved compatibility and parity between these components via standardization on exchange formats and alignment of APIs. The community plans to remove deprecated APIs and reduce the amount of duplication, which has caused confusion for users. Other improvements in TensorFlow 2.0 Increased Compatibility and continuity TensorFlow 2.0  would be an opportunity to correct mistakes and to make improvements which are otherwise restricted under semantic versioning. The community plans to create a conversion tool which updates the Python code to use TensorFlow 2.0 compatible APIs, to ease the transition for users. This tool will also warn in cases where conversion is not possible automatically. A similar tool helped tremendously during the transition to 1.0. As not all changes can be made fully, automatically, the community plans to deprecate APIs, some of which do not have a direct equivalent. For such cases, they will offer a compatibility module (tensorflow.compat.v1) which contains the full TensorFlow 1.x API, and will be maintained through the lifetime of TensorFlow 2.x. On-disk compatibility The community would not be making any breaking changes to SavedModels or stored GraphDefs repositories. This means they will include all current kernels in 2.0 (i.e., we plan to include all current kernels in 2.0). However, the changes in 2.0 will mean that variable names in raw checkpoints might have to be converted before being compatible with new models. Improvements to tf.contrib As part of releasing TensorFlow 2.0, the community will stop distributing tf.contrib. For each of the contrib modules they plan to  either: integrate the project into TensorFlow, move it to a separate repository, or remove it entirely. This means that all of tf.contrib will be deprecated, and the community will stop adding new tf.contrib projects. Following is a YouTube video by Aurélien Géron explaining the changes in TensorFlow 2.0 in detail. https://www.youtube.com/watch?v=WTNH0tcscqo Understanding the TensorFlow data model [Tutorial] TensorFlow announces TensorFlow Data Validation (TFDV) to automate and scale data analysis, validation, and monitoring Intelligent mobile projects with TensorFlow: Build your first Reinforcement Learning model on Raspberry Pi [Tutorial]
Read more
  • 0
  • 0
  • 3127