Lidar, which is like radar but with lasers instead of radio waves, can provide high-detail surveys of the land. The state of Washington is using the tool for beautiful results.
If you’re looking for love, start by cleaning your bathroom. This advice may seem counterintuitive since, let’s be real, romantic pursuits are not typically at the mercy of commode cleanliness (we hope). But, according to hundreds of thousands of people in the US on the dating app Hater, dirty bathrooms are quite literally, the worst.
The new-age dating app is pretty simple: users react (with love or hate) to thousands of topics, like people who don’t signal, beards, and Beyoncé. The result is a matchmaking service based on mutual likes and dislikes. So, if you’re a hater of dirty bathrooms, you’re in good company. In fact, here are the things that we found are the most universally loved and hated throughout the US.
University Data Science News
Hanna Wallach of Microsoft Research was a co-chair of NIPS where keynote speakerKate Crawford helped introduce a new focus on fairness, accuracy and transparency.
Karen Kelsky launched hashtag #MeTooPhD and an anonymous survey of people (1500 so far) in academia who have experienced sexual harassment. No names are named, but the survey results are public and it’s possible to search by institution and field. The upshot? Many women left academia altogether after being harassed by their supervisors and others have taken years to recover.
The ACM has a new report on the state of autonomous vehicles, explaining that simulations of driving conditions are required to improve the technology’s ability to respond correctly to predictably unpredictable events like plastic bags blowing across the road, deer jumping in front of vehicles, and icy conditions that simultaneously cause glare and demand extra care.
Yale University has established an annual Day of Data. This year the “Day” was December 1, marking the fifth year for the event. Based on Yale’s own description the event tilts more toward “tell” than “show.”
When Timnit Gebru was a student at Stanford University’s prestigious Artificial Intelligence Lab, she ran a project that used Google Street View images of cars to determine the demographic makeup of towns and cities across the U.S. While the AI algorithms did a credible job of predicting income levels and political leanings in a given area, Gebru says her work was susceptible to bias—racial, gender, socio-economic. She was also horrified by a ProPublica report that found a computer program widely used to predict whether a criminal will re-offend discriminated against people of color.
So earlier this year, Gebru, 34, joined a Microsoft Corp. team called FATE—for Fairness, Accountability, Transparency and Ethics in AI. The program was set up three years ago to ferret out biases that creep into AI data and can skew results.
The podcast gives Microsoft researchers a platform to share their work with a broad and curious audience, and aims for the perfect balance of professional and personal; technical and human. While the Microsoft Research Podcast is, first and foremost, about the research, it’s also an introduction to the researchers behind the projects. We find out who they are, what makes them tick, what inspires them, what gets them up in the morning, and what keeps them up at night. It’s not just the caffeine.
Every week, I talk to a new researcher and discover the stories behind the technology. The conversations are purposeful, but not scripted. Each interview reflects the unique nature of the subject matter and the person who has dedicated his or her life to it. Each interview reveals something surprising and amazing and each interview has a distinctly different flavor or theme.
Kellogg Insight, Based on the research of Joao Guerreiro, Sergio Rebelo and Pedro Teles
from
Earlier this year, Microsoft founder Bill Gates threw his support behind a controversial policy: a robot tax. As workers in many sectors are replaced by machines, the government is losing huge amounts of income tax revenue. Taxing the companies that employ robots, Gates reasoned, could help slow the pace of automation, and the revenue could be used to retrain employees.
Sergio Rebelo, a finance professor at the Kellogg School, had serious doubts when he heard Gates’s argument. For decades, economists have known that taxing so-called “intermediate goods”—goods that are used to make other goods, like the bricks used to build a house, or the robots used to manufacture cars—can make it harder for suppliers to create and sell their products.
“When you do that, you reduce the level of production in the economy,” Rebelo says.
“Machine Learning and AI experts from around the world will gather in Long Beach, CA, next week at NIPS 2017 to present the latest advances in machine learning and computational neuroscience. Research from Facebook will be presented in 10 peer-reviewed publications and posters.”
arXiv, Computer Science > Learning; Joel Hestness, Sharan Narang, Newsha Ardalani, Gregory Diamos, Heewoo Jun, Hassan Kianinejad, Md. Mostofa Ali Patwary, Yang Yang, Yanqi Zhou
from
Deep learning (DL) creates impactful advances following a virtuous recipe: model architecture search, creating large training data sets, and scaling computation. It is widely believed that growing training sets and models should improve accuracy and result in better products. As DL application domains grow, we would like a deeper understanding of the relationships between training set size, computational scale, and model accuracy improvements to advance the state-of-the-art.
This paper presents a large scale empirical characterization of generalization error and model size growth as training sets grow. We introduce a methodology for this measurement and test four machine learning domains: machine translation, language modeling, image processing, and speech recognition. Our empirical results show power-law generalization error scaling across a breadth of factors, resulting in power-law exponents—the “steepness” of the learning curve—yet to be explained by theoretical work. Further, model improvements only shift the error but do not appear to affect the power-law exponent. We also show that model size scales sublinearly with data size. These scaling relationships have significant implications on deep learning research, practice, and systems. They can assist model debugging, setting accuracy targets, and decisions about data set growth. They can also guide computing system design and underscore the importance of continued computational scaling.
So where’s the opportunity for startups? At Homebrew, we’ve focused on ML/AI investing in companies that have two key elements: 1) data that is largely hidden from the incumbents and 2) cost effective, scalable win-the-market strategies (plain old go-to-market ain’t good enough!). Our view is the combination of these two elements leads to long term advantage and a competitive moat that protects against both the incumbents and other startups that enter the market.
Finding data pools that are black boxes to incumbents usually means targeting verticals or use cases where the larger players don’t have access, experience or focus.
Alexandria, VAProfessor Juan de Pablo from the Institute for Molecular Engineering at University of Chicago will be giving a lecture on Soft Materials Research in the Era of Machine Learning at the National Science Foundation on Monday, December 11 at 2:00 p.m.
Peyton is a Python library that allows you to interact with the Throne.ai platform for sports prediction. From this library, you can download historical and competition data, user ranking data, and submit predictions.
“Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to find insights and relationships in text. Amazon Comprehend identifies the language of the text; extracts key phrases, places, people, brands, or events; understands how positive or negative the text is; and automatically organizes a collection of text files by topic.”
“Binder allows researchers to quickly create the computational environment needed to interact with research code and data shared online. To interact with someone else’s work, you simply click a URL and are taken directly to a live environment where you can run the code in the cloud.”
“AWS today launched a new browser-based IDE, AWS Cloud9, at its re:Invent conference in Las Vegas. Cloud9 is actually an IDE the company acquired last year. It’s based on the Ace Editor and the work Cloud9 did on its c9.io IDE.”