Before you choose to head back to the Classic look of the site, we'd appreciate it if you share your thoughts on the Beta; your feedback is what drives our ongoing development.
Beta is different and we value you taking the time to try it out. Please take a look at the changes we've made in Beta and learn more about it. Thanks for reading, and for making the site better!
An anonymous reader writes: Brianna Wu, leader of a game development studio, has an article exposing the constant harassment of women in the games industry. She says, "I'm not writing this piece to evoke your sympathy. I'm writing to share with you what prominent, successful women in the industry experience, in their own words." She goes through the individual stories of several women targeted by this vitriol, and tries to figure out why it happens. Quoting: "We live in a society that's sexist in ways it doesn't understand. One of the consequences is that men are extremely sensitive to being criticized by women. ... This is why women are socialized to carefully dance around these issues, disagreeing with men in an extremely gentle manner. Not because women are nicer creatures than men. But because our very survival can depend on it. ... Growing a thicker skin isn't the answer, nor is it a proper response. Listening, and making the industry safer for the existence of visible women is the best, and only, way forward."
rjmarvin writes: Microsoft Research is testing a new method for predicting errors and bugs while developers write code: biometrics. By measuring a developer's eye movements, physical and mental characteristics as they code, the researchers tracked alertness and stress levels to predict the difficulty of a given task with respect to the coder's abilities. In a paper entitled "Using Psycho-Physiological Measures to Assess Task Difficulty in Software Development," the researchers summarized how they strapped an eye tracker, an electrodermal sensor and an EEG sensor to 15 developers as they programmed for various tasks. Biometrics predicted task difficulty for a new developer 64.99% of the time. For a subsequent tasks with the same developer, the researchers found biometrics to be 84.38% accurate. They suggest using the information to mark places in code that developers find particularly difficult, and then reviewing or refactoring those sections later.
dcblogs (1096431) writes On the floor of U.S. Senate Thursday, Sen. Jeff Sessions delivered a scalding and sarcastic attack on the use of highly skilled foreign workers by U.S. corporations that was heavily aimed at Microsoft, a chief supporter of the practice. Sessions' speech began as a rebuttal to a recent New York Times op-ed column by Microsoft founder Bill Gates, investor Warren Buffett and Sheldon Adelson ... But the senator's attack on "three of our greatest masters of the universe," and "super billionaires," was clearly primed by Microsoft's announcement, also on Thursday, that it was laying off 18,000 employees. "What did we see in the newspaper today?" said Sessions, "News from Microsoft. Was it that they are having to raise wages to try to get enough good, quality engineers to do the work? Are they expanding or are they hiring? No, that is not what the news was, unfortunately. Not at all."
An anonymous reader writes: There's often debate amongst modern programmers about how much math a professional developer should know, and to what extent programming is math. Learning to program is often viewed as being on a spectrum between learning math and learning spoken/written languages. But in a new article, Jeremy Kun argues that the spectrum should be formulated another way: Human language -> Mathematics -> Programming. "Having studied all three subjects, I'd argue that mathematics falls between language and programming on the hierarchy of rigor. ... [T]he hierarchy of abstraction is the exact reverse, with programming being the most concrete and language being the most abstract. Perhaps this is why people consider mathematics a bridge between human language and programming. Because it allows you to express more formal ideas in a more concrete language, without making you worry about such specific hardware details like whether your integers are capped at 32 bits or 64. Indeed, if you think that the core of programming is expressing abstract ideas in a concrete language, then this makes a lot of sense. This is precisely why learning mathematics is 'better' at helping you learn the kind of abstract thinking you want for programming than language. Because mathematics is closer to programming on the hierarchy. It helps even more that mathematics and programming readily share topics."
New submitter gthuang88 (3752041) writes In the 1990s, Microsoft was in position to own the software and devices market. Here is Nathan Myhrvold's previously unpublished 1997 memo on expanding Microsoft Research to tackle problems in software testing, operating systems, artificial intelligence, and applications. Those fields would become crucial in the company's competition with Google, Apple, Amazon, and Oracle. But research didn't do enough to make the company broaden its businesses. While Microsoft Research was originally founded to ensure the company's future, the organization only mapped out some possible futures. And now Microsoft is undergoing the biggest restructuring in its history. At least F# and LINQ saw the light of day.
ectoman writes: Opensource.com is running an interview with Jennifer Davidson of ChickTech, a non-profit organization whose mission is to create communities of support for women and girls pursuing (or interested in pursuing) careers in tech. "In the United States, many girls are brought up to believe that 'girls can't do math' and that science and other 'geeky' topics are for boys," Davidson said. "We break down that idea." Portland, OR-based ChickTech is quickly expanding throughout the United States—to cities like Corvallis and San Francisco—thanks to the "ChickTech: High School" initiative, which gathers hundreds of young women for two-day workshops featuring open source technologies. "We fill a university engineering department with 100 high school girls—more girls than many engineering departments have ever seen," Davidson said. "The participants can look around the building and see that girls from all backgrounds are just as excited about tech as they are."
An anonymous reader writes with some exciting news from the world of processor design: Robert Watson at Cambridge (author of Capsicum) has written a blog post on SRI/Cambridge's recent open sourcing of the hardware and software for the DARPA-sponsored CHERI processor — including laser cutting directions for an FPGA-based tablet! Described in their paper The CHERI Capability Model: Reducing Risk in an age of RISC, CHERI is a 64-bit RISC processor able to boot and run FreeBSD and open-source applications, but has a Clang/LLVM-managed fine-grained, capability-based memory protection model within each UNIX process. Drawing on ideas from Capsicum, they also support fine-grained in-process sandboxing using capabilities. The conference talk was presented on a CHERI tablet running CheriBSD, with a video of the talk by student Jonathan Woodruff (slides).
Although based on the 64-bit MIPS ISA, the authors suggest that it would also be usable with other RISC ISAs such as RISC-V and ARMv8. The paper compares the approach with several other research approaches and Intel's forthcoming Memory Protection eXtensions (MPX) with favorable performance and stronger protection properties. The processor "source code" (written in Bluespec Verilog) is available under a variant of the Apache license (modified for application to hardware). Update: 07/16 20:53 GMT by U L : If you have any questions about the project, regular Slashdot contributor TheRaven64 is one of the authors of the paper, and is answering questions.
An anonymous reader writes Writer and former software engineer Matt Gemmell adds his voice to the recent rumblings about writing code as a profession. Gemmell worries that the latest "software Renaissance," which was precipitated by the explosion of mobile devices, is drawing to a close. "Small shops are closing. Three-person companies are dropping back to sole proprietorships all over the place. Products are being acquired every week, usually just for their development teams, and then discarded. The implacable, crushing wheels of industry, slow to move because of their size, have at last arrived on the frontier. Our frontier, or at least yours now. I've relinquished my claim." He also pointed out the cumulative and intractable harm being done by software patents, walled-garden app stores, an increasingly crowded market, and race-to-the-bottom pricing. He says that while the available tools make it a fantastic time to develop software, actually being an independent developer may be less sustainable than ever.
An anonymous reader writes: My niece, who is graduating from high school, has asked me for some career advice. Since I work in data processing, my first thought was to recommend a degree course in computer science or computer engineering. However, after reading books by Jeremy Rifkin (The Third Industrial Revolution) and Ray Kurzweil (How to Create a Mind), I now wonder whether a career in information technology is actually better than, say, becoming a lawyer or a construction worker. While the two authors differ in their political persuasions (Rifkin is a Green leftist and Kurzweil is a Libertarian transhumanist), both foresee an increasingly automated future where most of humanity would become either jobless or underemployed by the middle of the century. While robots take over the production of consumer hardware, Big Data algorithms like the ones used by Google and IBM appear to be displacing even white collar tech workers. How long before the only ones left on the payroll are the few "rockstar" programmers and administrators needed to maintain the system? Besides politics and drug dealing, what jobs are really future-proof? Would it be better if my niece took a course in the Arts, since creativity is looking to be one of humanity's final frontiers against the inevitable Rise of the Machines?
sandbagger (654585) writes with word of a Y2K-style bug showing up in Y2K14: "The glitch originated with the Pennsylvania Department of Motor Vehicles during an automated data transfer of nearly 400,000 records. The records of males born between 1993 and 1997 were mixed with those of men born a century earlier. The federal agency didn't know it because the state uses a two-digit code to indicate birth year." I wonder where else two-digit years are causing problems; I still see lots of paper forms that haven't made the leap yet to four digits.
When Linus Torvalds first announced his new operating system project ("just a hobby, won't be big and professional like gnu"), he aimed the announcement at users of Minix for a good reason: Minix (you can download the latest from the Minix home page) was the kind of OS that tinkerers could afford to look at, and it was intended as an educational tool. Minix's creator, Professor Andrew Stuart "Andy" Tanenbaum, described his academic-oriented microkernel OS as a hobby, too, in the now-famous online discussion with Linus and others. New submitter Thijssss (655388) writes with word that Tanenbaum, whose educational endeavors led indirectly to the birth of Linux, is finally retiring. "He has been at the Vrije Universiteit for 43 years, but everything must eventually end."
An anonymous reader writes "I'm a Solaris user which is not well supported by the OSS toolchains. I'd like to have a dedicated Linux based dev system which has good support for ARM, MSP430 and other MCU lines and draws very little (5-10 watts max) power. The Beaglebone Black has been suggested. Is there a better choice? This would only be used for software development and testing for embedded systems."
theodp (442580) writes Over at Alarming Development, Jonathan Edwards has an interesting rant entitled Developer Inequality and the Technical Debt Crisis. The heated complaints that the culture of programming unfairly excludes some groups, Edwards feels, is a distraction from a bigger issue with far greater importance to society.
"The bigger injustice," Edwards writes, "is that programming has become an elite: a vocation requiring rare talents, grueling training, and total dedication. The way things are today if you want to be a programmer you had best be someone like me on the autism spectrum who has spent their entire life mastering vast realms of arcane knowledge — and enjoys it. Normal humans are effectively excluded from developing software. The real injustice of developer inequality is that it doesn't have to be this way." Edwards concludes with a call to action, "The web triumphalists love to talk about changing the world. Well if you really want to change the world, empower regular people to build web apps. Disrupt web programming! Who's with me?" Ed Finkler, who worries about his own future as a developer in The Developer's Dystopian Future, seconds that emotion. "I think about how I used to fill my time with coding," Finkler writes. "So much coding. I was willing to dive so deep into a library or framework or technology to learn it. My tolerance for learning curves grows smaller every day. New technologies, once exciting for the sake of newness, now seem like hassles. I'm less and less tolerant of hokey marketing filled with superlatives. I value stability and clarity."
ShaunC writes: Is there a glut of qualified American tech workers, or isn't there? Some companies like Facebook and Airbnb are now actively courting and recruiting high school students as young as 13 with promises of huge stipends and salaries. As one student put it, "It's kind of insane that you can make more than the U.S. average income in a summer." Another who attended a Facebook-sponsored trip said he'd "forego college for a full-time job" if it were offered. Is Silicon Valley taking advantage of naive young workers?
itwbennett writes: Python has surpassed Java as the top language used to introduce U.S. students to programming and computer science, according to a recent survey posted by the Association for Computing Machinery (ACM). Eight of the top 10 computer science departments now use Python to teach coding, as well as 27 of the top 39 schools, indicating that it is the most popular language for teaching introductory computer science courses, according to Philip Guo, a computer science researcher who compiled the survey for ACM."
Nerval's Lobster (2598977) writes Technology writer Jon Brodkin sat down with a group of indie game developers (as well as a professor at the University of Southern California's game-design program) to talk about why they decided to launch their own small studios rather than stick with comfortable (albeit stressful) jobs at major firms like Disney or Zynga. The answer, as you'd expect, boils down to control. "Working for a bigger company is a good way to gain experience, and learn how games are made," said Graham Smith, one of the co-founders of Toronto-based DrinkBox Studios. "It's also nice to have a steady salary coming in as you learn the ropes. On the flip side, depending on the company, you might not have much control over the game's design, or even be making the types of games that you enjoy playing." But startups come with their own challenges, not the least of which is the prospect of an economic downturn quickly wiping you out, or not making your Kickstarter goal.
itwbennett (1594911) writes "How do you measure success? If it's by Stack Overflow reputation, Google engineer Jon Skeet is the world's best programmer. If it's winning programming competitions, Gennady Korotkevich or Petr Mitrechev might be your pick. But what about Linus Torvalds? Or Richard Stallman? Or Donald Knuth? ITworld's Phil Johnson has rounded up a list of what just might be the world's top 14 programmers alive today."
KDE Community (3396057) writes The KDE Community is proud to announce the release of KDE Frameworks 5.0. Frameworks 5 is the next generation of KDE libraries, modularized and optimized for easy integration in Qt applications. The Frameworks offer a wide variety of commonly needed functionality in mature, peer reviewed and well tested libraries with friendly licensing terms. There are over 50 different Frameworks as part of this release providing solutions including hardware integration, file format support, additional widgets, plotting functions, spell checking and more. Many of the Frameworks are cross platform and have minimal or no extra dependencies making them easy to build and add to any Qt application. Version five of the desktop shell, Plasma, will be released soon, and packages of Plasma-next and KDE Frameworks 5 will trickle into Ubuntu Utopic over the next few days. There's a Live CD of Frameworks 5 / Plasma-next, last updated July 4th.
An anonymous reader writes with news about how Oculus is dealing with the reselling of dev kits in China. Bad news for those of you looking to get your hands on a preorder of the Oculus Rift virtual reality headset. A representative from Oculus recently confirmed that the company has had to stop selling its headsets in China as a result of an undisclosed amount of reselling. Which is to say, some of those preordering the developer edition of the virtual reality headset in China — not the consumer product, which hasn't been officially released in any capacity just yet — aren't actually looking to develop anything on the headsets. Nor are they even interested in getting a first look at the virtual reality capabilities of the $350 development kit. They're scalping, plain and simple, to take advantage of what appears to be a hefty amount of demand for the device.
An anonymous reader writes Working with computational journalist Nick Diakopoulos, we at IEEE Spectrum have published an app that ranks the popularity of dozens of programming languages. Because different fields have different interests (what's popular with programmers writing embedded code versus what's hot with web developers isn't going to be identical) we tried to make the ranking system as transparent as possible — you can use our presets or you can go in and create your own customized ranking by adjusting the individual weightings of the various data sources we mined.
An anonymous reader writes "We all know somebody who changes jobs like changing clothes. In software development and IT, it's getting increasingly hard to find people who have been at their job for more than a few years. That's partly because of tech companies' bias for a young work force, and partly because talented people can write their own ticket in this industry. Thus, I put the question to you: how often should you be switching jobs? Obviously, if you find the perfect company (full of good people, doing interesting things, paying you well), your best bet is to stay. But that's not the reality for most of the workforce. Should you always be keeping an eye out for new jobs? Is there a length of time you should stick around so you don't look like a serial job-hopper? Does there come a point in life when it's best to settle down and stick with a job long term?"
theodp writes:Google recently announced Global Impact Awards for Computer Science, part of the company's $50 million investment to get girls to code. But Google's influence over K-12 CS education doesn't stop there. The Sun-Times reports that Chicago Public School (CPS) teachers are participating in a summer professional development program hosted by Google as part of the district's efforts to "saturate" schools with CS within 3 years: "The launch of CS4All [Computer Science for All], in partnership with Code.org and supported by Google, starts this fall in 60 CPS schools to try to bridge the digital divide and prepare students." And in two weeks, the Computer Science Teachers Association [CSTA] and Google will be presenting the National Computer Science Principles Education Summit. "Attendees at this event have been selected through a rigorous application process that will result in more than 70 educators and administrators working together to strategize about getting this new Advanced Placement course implemented in schools across the country," explains CSTA. The ACM, NSF, Google, CSTA, Microsoft, and NCWIT worked together in the past "to provide a wide range of information and guidance that would inform and shape CS education efforts," according to the University of Chicago, which notes it's now conducting a follow-up NSF-funded study — Barriers and Supports to Implementing Computer Science — that's advised by CPS, CSTA, and Code.org.
Famo.us chases another holy grail, namely the 'write once, run anywhere' dream. Instead of having to write different code for different platforms, like iOS and Android, developers can write one application that works and looks as good on all platforms, in theory anyway. This of course saves a huge amount of time and resources. Unfortunately, this idea is not without its problems and has never really worked very well with earlier attempts like Java-applets, Flash and Silverlight. In the end native applications have so far always been faster and slicker and I'm pretty skeptical Famo.us will be able to change this."
M-Saunders writes: Perl 6 has been in development since 2000. So why, 14 years later, hasn't it been released yet? Linux Voice caught up with Damian Conway, one of the architects of Perl 6, to find out what's happening. "Perl 6 has all of the same features [as Perl 5] but with the rough edges knocked off of them", he says. Conway also talks about the UK's Year of Code project, and how to get more people interested in programming.
An anonymous reader writes "I work for a huge company. We use SAS all the time for everything, which is great if you have a bunch of non-programmer employees and you want them to do data analysis and build models... but it ends up stifling any real innovation, and I worry we will get left behind. Python and R both seem to be emerging stars in the data science game, so I would like to steer us towards one of them. What compelling arguments can you give that would help an old company change its standard if that company is pretty set in its ways?"
MojoKid writes One of the trickiest aspects to launching a new platform update is the chicken and egg problem. Without any hardware to test on, developers are leery of committing to supporting new hardware features. Without software that takes advantage of new hardware capabilities, customers aren't willing to pay for new equipment. This is the crux of the issue with respect to the ARMv8 architecture and enabling development for 64-bit Android platforms. As such ARM is readying their Juno development platform that combines several of ARM's most advanced technologies on a single board. The product supports big.Little in an asymmetric configuration; each board ships with two Cortex-A57s, four Cortex-A53s, and a modest Mali T-624 core. All this hardware needs an OS to run on — which is why ARM is announcing a 64-bit port of Android as part of this new development board. By including AOSP support as well as additional hooks and features from Linaro, ARM wants Juno to be a sort-of one-stop shopping product for anyone who needs to test, prototype, or design a 64-bit product for the ARM ecosystem. The Android flavor that's coming over is based on Linaro Stable Kernel 3.10. At launch, Juno will support OpenGL-ES 3.0, on-chip thermal and power management, up to 8GB of RAM (12.8GB/s of bandwidth), an optional FPGA, and USB 2.0. OpenCL 1.1 will be added in a future product update. The project is positioned as a joint ARM / Linaro launch with ARM handling the hardware and Linaro taking responsibility for the software stack.
adelayde (185757) writes "In my day job, I work on a web based service with a lot of legacy code written in that older (and some may say venerable) web-scripting language, Perl. Although we use Modern Perl extensions such as Moose, the language just seems to be ossifying and we're wanting to move to a more up-to-date and used language for web applications, or even an entire framework, to do new development. We're still planning to support the legacy code for a number of years to come; that's unavoidable. This is a fairly big project and it's mission critical to the business. The thing we're afraid of is jumping onto something that is too new and too buzzy as we'd like to make a technology decision that would be good at least for the next five years, if not more, and today's rising star could quite easily be in tomorrow's dustbin. What language and/or framework would you recommend we adopt?"
First time accepted submitter HagraBiscuit (2756527) writes Researchers from the Faculty of Computer Science, Free University of Bozen-Bolzano, Bolzano, Italy, have quantified and analysed affective mood index against objective measures of problem-solving effectiveness for a group of software developers. From the report abstract: "The results offer support for the claim that happy developers are indeed better problem solvers in terms of their analytical abilities. The following contributions are made by this study: (1) providing a better understanding of the impact of affective states on the creativity and analytical problem-solving capacities of developers, (2) introducing and validating psychological measurements, theories, and concepts of affective states, creativity, and analytical-problem-solving skills in empirical software engineering, and (3) raising the need for studying the human factors of software engineering by employing a multidisciplinary viewpoint.
theodp (442580) writes In an interview with The Washington Post's Lyndsey Layton that accompanied her report on How Bill Gates Pulled Off the Swift Common Core Revolution (the Gates Foundation doled out $233 million in grants to git-r-done), Gates denied that he has too much influence in K-12 education. Despite Gates' best efforts, however, there's been more and more pushback recently from both teachers and politicians on the standards, GeekWire's Taylor Soper reports, including a protest Friday by the Badass Teacher Association, who say Gates is ruining education. "We want to get corporations out of teaching," explained one protester. If that's the case, the "Badasses" probably won't be too pleased to see how the K-12 CS education revolution is shaping up, fueled by a deep-pocketed alliance of Gates, Google, Microsoft, Facebook, and others. Google alone has already committed $90 million to influence CS education. And well-connected Code.org, which has struck partnerships with school districts reaching over 2M U.S. students and is advising NSF-funded research related to the nation's CS 10K Project, will be conducting required professional development sessions for K-12 CS teachers out of Google, Microsoft, and Amazon offices this summer in Chicago, New York City, Boston, and Seattle. So, could K-12 CS Education ("Common Code"?) become the next Common Core?
redletterdave writes: According to a blog post from Gregg Pollack, CEO of the Code School, Google is paying for three free months for any women and minorities interested in tech to expand their skills. The offer is part of Google's $50 million "Made With Code" initiative, which aims to help close the gender gap in tech. While Google is also offering the same vouchers to the women in attendance at its annual I/O developers conference this week, the search giant has released an online application that's available to women everywhere. Google says its available vouchers for women number in the "thousands."
tyggna writes: "The flame wars of different shells and text editors have long been established, but my question is this: are text editors and various languages linked? Do the majority of Ruby programmers use Emacs? Are most Perl programmers using vim?
Please post your editor and language of choice in the comments."
shocking writes: Arizona National Guard member Vivin Paliath was surprised to be commended for writing Perl scripts and Excel macros while his unit was deployed in Iraq. His work automated a number of previously manual processes that were part of the logistics processes of his unit. He wrote, '[A]s a programmer, I'm constantly looking for ways to make my job easy. I didn't want to sit and add qualifications, and print licenses one by one. I was too lazy for that, and worse, the whole thing was horribly inefficient. So I decided to figure out how to automate the process. ... I started writing Perl scripts to query the data. By the time we had reached Iraq, I had a working script that generated licenses as text files for all the soldiers. The script only took a second or two to run, and the longest part of the process was simply printing out the licenses. But I wasn't done yet. I was still annoyed that I would have to add driver qualifications manually. So I wrote another script that would go and add qualifications to drivers en masse. The script even had a configuration file where you could specify what qualifications you wanted to add and to whom."
An anonymous reader writes "Many people reading this site probably have a functional understanding of how algorithms work. But whether you know algorithms down to highly mathematical abstractions or simple as a fuzzy series of steps that transform input into output, it can be helpful to visualize what's going on under the hood. That's what Mike Bostock has done in a new article. He walks through algorithms for sampling, shuffling, and maze generation, using beautiful and fascinating visualizations to show how each algorithm works and how it differs from other options.
He says, "I find watching algorithms endlessly fascinating, even mesmerizing. Particularly so when randomness is involved. ... Being able to see what your code is doing can boost productivity. Visualization does not supplant the need for tests, but tests are useful primarily for detecting failure and not explaining it. Visualization can also discover unexpected behavior in your implementation, even when the output looks correct. ...Even if you just want to learn for yourself, visualization can be a great way to gain deep understanding. Teaching is one of the most effective ways of learning, and implementing a visualization is like teaching yourself."
An anonymous reader writes: Hello Games is a small development studio, only employing 10 people. But they're building a game, No Man's Sky, that's enormous — effectively infinite. Its universe is procedurally generated, from the star systems down to individual species of plant and animal life. The engine running the game is impressively optimized. A planet's characteristics are not computed ahead of time — terrain and lifeforms are randomly generated on the fly as a player explores it. But, of course, that created a problem for the developers — how do they know their procedural generation algorithms don't create ridiculous life forms or geological formations? They solved that by writing AI bot software that explores the universe and captures brief videos, which are then converted to GIF format and posted on a feed the developers can review. The article goes into a bit more detail on how the procedural generation works, and how such a small studio can build such a big game.
itwbennett writes: A group of researchers from Google, the Hong Kong University of Science and Technology and the University of Nebraska undertook a study of over 26 million builds by 18,000 Google engineers from November 2012 through July 2013 to better understand what causes software builds to fail and, by extension, to improve developer productivity. And, while Google isn't representative of every developer everywhere, there are a few findings that stand out: Build frequency and developer (in)experience don't affect failure rates, most build errors are dependency-related, and C++ generates more build errors than Java (but they're easier to fix).
Google I/O, the company's annual developer tracking^wdevelopers conference, has opened today in San Francisco. This year the company has reduced the number of conference sessions to 80, but also promised a broader approach than in previous years -- in other words, there may be a shift in focus a bit from Google's best known platforms (Chrome/Chrome OS and Android). Given its wide-ranging acquisitions and projects (like the recent purchase of Nest, which itself promptly bought Dropcam, the ever smarter fleet of self-driving cars, the growing number of Glass devices in the wild, and the announcement of a 3D scanning high end tablet quite unlike the Nexus line of tablets and phones), there's no shortage of edges to focus on. Judging from the booths set up in advance of the opening (like one with a sign announcing "The Physical Web," expect some of the stuff that gets lumped into "the Internet of Things." Watch this space -- updates will appear below -- for notes from the opening keynote, or follow along yourself with the live stream, and add your own commentary in the comments. In the days to come, watch for some video highlights of projects on display at I/O, too. Update: 06/25 17:41 GMT by T : Updates rolling in below on Android, wearables, Android in cars, Chromecast, smart watches, etc.Keep checking back! (Every few minutes, I get another chunk in there.)
mpicpp writes "Today, in advance of Google I/O, Nest has officially announced a new developer program and API that will allow other companies' smart devices to communicate with Nest's Protect smoke alarm and Learning Thermostat. Among the companies that Nest is partnering with for this initial publicity push are IFTTT, Jawbone, LIFX, Logitech, Mercedes-Benz, Whirlpool, Chamberlain, and Google itself—the latter two companies will release Nest-compatible features this fall, while the others are all available today.
An anonymous reader writes: Here's a fun project: engineer Yuriy Guts built a Visual Studio extension that lets people program using MIDI instruments. You can write code letter by letter on a piano keyboard. Granted, it's not terribly efficient, but it's at least artistic — you can compose music that is also a valid computer program. Somewhat more usefully, it also allows you to turn a simple MIDI input device, like a trigger pad into a set of buttons that will run tests, push/pull code, or other tasks suitable for automation. The extension is open source and open to contributions.
mpicpp (3454017) writes with word that Mozilla released a full development environment integrated into Firefox (available now in nightly builds). From the announcement: Developers tell us that they are not sure how to start app development on the Web, with so many different tools and templates that they need to download from a variety of different sources. We’re solving that problem with WebIDE, built directly into Firefox. Instead of starting from zero we provide you with a functioning blueprint app with the click of a button. You then have all the tools you need to start creating your own app based on a solid foundation. WebIDE helps you create, edit, and test a new Web application right from your browser. It lets you install and test apps on Firefox OS devices and simulators and integrates the Firefox Developer Tools for seamless debugging and inspection across those devices. This is a first step towards debugging across various platforms and devices over WiFi using open remote debugging APIs. The default editor is based on CodeMirror, but the protocol for interacting with the IDE is open and support for other editors (Emacs anyone?) should appear soon.
An anonymous reader writes Oracle is buying hospitality and retail technology vendor Micros Systems for $5.3 billion, in a deal that will be its largest since the purchase of Sun Microsystems in 2010. "Oracle said the acquisition will extend its offerings by combining Micros' industry-specific applications with its business applications, technologies and cloud portfolio. Oracle expects the deal to immediately add to its adjusted earnings. Its stock climbed 18 cents to $41 before the market opened. Micros' board unanimously approved the transaction, which is expected to close in the second half of the year."
Presto Vivace writes: Fortune has an article about increasingly overt age discrimination in the tech industry. Quoting: "It's a widely accepted reality within the technology industry that youth rules. But at least part of the extreme age imbalance can be traced back to advertisements for open positions that government regulators say may illegally discriminate against older applicants. Many tech companies post openings exclusively for new or recent college graduates, a pool of candidates that is overwhelmingly in its early twenties. ... 'In our view, it's illegal,' Raymond Peeler, senior attorney advisor at the Equal Employment Opportunity Commission, the federal agency that enforces workplace discrimination laws said about the use of 'new grad' and 'recent grad' in job notices. 'We think it deters older applicants from applying.'" Am I the only one who thinks many of the quality control issues and failed projects in the tech industry can be attributed to age discrimination?
An anonymous reader writes So I, like many people, want to make my own game. Outside of MATLAB, Visual Basic, and LabVIEW I have no real programming experience. I initially started with Ruby, but after doing my homework decided that if I ever wanted to progress to a game that required some power, I would basically need to learn some form of C anyway. Further digging has led me to C#. The other parts of game design and theory I have covered: I have ~8 years of CAD modeling experience including Maya and Blender; I have a semiprofessional sound studio, an idie album on iTunes, and am adept at creating sound effects/music in a wide variety of programs; I'm familiar with the setbacks and frustration involved with game development — I beta tested DotA for 9ish years; I already have my game idea down on paper (RTS), including growth tables, unit types, unit states, story-lines, etc. I've been planning this out for a year or two; I will be doing this on my own time, by myself, and am prepared for it to take a couple years to finish. The reason for listing that stuff out, is that I want people to understand that I know what I'm getting myself in to, and I'm not trying to put out a not-so-subtle "help me make a game for free lol" type of post. With all of that said, where is a good place to start (i.e., recommended books) for learning C# for game programming? I am familiar with object oriented programming, so that's a little bit of help. I'm not necessarily looking for the syntax (that part is just memorization), but more for the methodology involved. If anyone also has any suggestions for other books or information that deal with game development, I would love to hear that too. I know enough to understand that I really don't know anything, but have a good foundation to build on.
jfruh writes: "Creators of compilers are in an arms race to improve performance. But according to a presentation at this week's annual USENIX conference, those performance boosts can undermine your code's security. For instance, a compiler might find a subroutine that checks a huge bound of memory beyond what's allocated to the program, decide it's an error, and eliminate it from the compiled machine code — even though it's a necessary defense against buffer overflow attacks."
theodp writes: On Thursday, Google announced a $50 million initiative to inspire girls to code called Made with Code. As part of the initiative, Google said it will also be "rewarding teachers who support girls who take CS courses on Codecademy or Khan Academy." The rewards are similar to earlier coding and STEM programs run by Code.org and Google that offered lower funding or no funding at all to teachers if participation by female students was deemed unacceptable to the sponsoring organizations. The announcement is all the more intriguing in light of a Google job posting seeking a K-12 Computer Science Education Outreach Program Manager to "work closely with external leaders and company executives to influence activities that drive toward collaborative efforts to achieve major 'moonshots' in education on a global scale." Perhaps towards that end, Google recently hired the Executive Director of the Computer Science Teachers Association (CSTA), who was coincidentally also a Code.org Advisory Board member. And Code.org — itself a Made With Code grantee — recently managed to lure away the ACM's Director of Public Policy to be its COO. So, are these kinds of private-public K-12 CS education initiatives (and associated NSF studies) a good idea? Some of the nation's leading CS educators sure seem to think so (video).
harrymcc (1641347) writes "In June 1967, a weekly newspaper called Computerworld launched. Almost exactly 47 years later, it's calling it quits in print form to focus on its website and other digital editions. The move isn't the least bit surprising, but it's also the end of an era--and I can' t think of any computing publication which had a longer run. Over at Technologizer, I shared some thoughts on what Computerworld meant to the world, to its publisher, IDG, and to me."
angry tapir (1463043) writes Third party developers will be able to build mobile applications that tap into the features of Adobe's Creative Cloud, including effects such as Photoshop's "content-aware fill" and PSD file manipulation, thanks to a new SDK the company is releasing as part of a major update to the suite of graphic design products. However, the company has been mum on important details such as how much (if anything) it will cost and what the license is likely to be (at the very least it seems end users will need to be Creative Cloud subscribers). The company has also made a foray into hardware releasing a pressure-sensitive stylus for tablets called Ink and a ruler called Slide.