Charting the progress towards AngularJS 2.0
AngularJS, for me, was a revelation the first time I encountered it. I was coming from using GWT (Google Web Toolkit), and seeing our large application shrink in lines of code by 90% was close to a spiritual experience. I was a convert from day one, because I knew how bad things were otherwise. Ever since, I have been associated with AngularJS in one way or another, and have seen how it makes things absolutely simple with data binding, templating, routing, unit testing, and so much more. But the more I used it, some things didn’t make sense, from naming to concepts. I got the hang of it, but I never really got to like why directives needed to be so complex, or how the built-in routing was quite limiting. While AngularJS made it trivial to write applications, it also made it trivial to write slow, hard-to-maintain applications if you didn’t understand how it all worked together.
Natalie Kuldell on the hard work of bringing biocoding to the classroom.
Synthetic biology is poised to change everything from energy development to food production to medicine — but there’s a bottleneck looming. How fast things develop depends on the number of people developing things. Let’s face it: there aren’t that many biocoders. Not in the universities, not in industry, not in the DIY sector. Not enough to change the world, at any rate. We have to ramp up.
And that means we first must train teachers and define biocoding curricula. Not at the university level — try secondary, maybe even primary schools. That, of course, is a challenge. To get kids interested in synthetic biology, we have to do just that: get them interested. More to the point, get them jacked. Biocoding is incredibly exciting stuff, but that message isn’t getting across.
“Students think science and engineering is removed from daily life,” says Natalie Kuldell, an instructor of biological engineering at MIT. “We have to get them engaged, and connected to science and engineering — more specifically, bioengineering — in meaningful ways.”
Network neutrality is about treating all kinds of traffic equally — throttling competition equates to extortion.
I’d like to make a few very brief points about net neutrality. For most readers of Radar, there’s probably nothing new here, but they address confusions that I’ve seen.
- Network neutrality isn’t about the bandwidth that Internet service providers deliver to your home. ISPs can charge more for more bandwidth, same as always.
- Nor is network neutrality about the bandwidth that Internet service providers deliver to information providers. Again, ISPs can charge more for more bandwidth, same as always. You’d better believe that Google pays a lot more for Internet service than your local online store.
- Nor is network neutrality about ISPs dealing with congestion. Network providers have always dealt with congestion — in the worst case, by dropping traffic. Remember the “fast busy” signal on the phone? That’s the network dealing with congestion.
- Network neutrality is entirely about treating all kinds of traffic equally. Video is the same as voice, the same as Facebook, the same as Amazon. Your ISP cannot penalize video traffic (or some other kind of traffic) because they’d like to get into that business or because they’re already in that business. In other words: when you buy Internet connectivity, you can use it for whatever you want. Your provider can’t tell you what kind of business to be in.
How a small and passionate team used modern techniques to shift a business on a short timeline.
Over the past year, I assisted in creating an application that helped shift a major part of IBM to a software-as-a-service (SaaS) model. I did this with the help of a small but excellent development team that was inspired by the culture and practices of web startups. To be clear, it wasn’t easy – changing how we worked led to frequent friction and conflict – but in the end it worked, and we made a difference.
In mid-2013, the IBM Service Management business and engineering leaders decided to make a big bet on moving our software to the cloud. Traditionally we have sold “on premises” software products. These are software products that a customer buys, downloads, and installs on their own equipment, in their own data centers and facilities. Although we love the on-premises business, we realized that cloud delivery of software is also a great option, and as our customers evolved to a hybrid on-premises / cloud future, we needed to be there to help them.
Google requires quid for its quo, but it offers something many don’t: user data access.
Despite some misgivings about the company’s product course and service permanence (I was an early and fanatical user of Google Wave), my relationship with Google is one of mutual symbiosis. Its “better mousetrap” approach to products and services, the width and breadth of online, mobile, and behind-the-scenes offerings saves me countless hours every week in exchange for a slice of my private life, laid bare before its algorithms and analyzed for marketing purposes.
I am writing this on a Chromebook by a lake, using Google Docs and images in Google Drive. I found my way here, through the thick underbrush along a long since forgotten former fishmonger’s trail, on Google Maps after Google Now offered me a glimpse of the place as one of the recommended local attractions.
Admittedly, having my documents, my photos, my to-do lists, contacts, and much more on Google, depending on it as a research tool and mail client, map provider and domain host, is scary. And as much as I understand my dependence on Google to carry the potential for problems, the fact remains that none of those dependencies, not one shred of data, and certainly not one iota of my private life, is known to the company without my explicit, active, consent. Read more…
How neuroscience is benefiting from distributed computing — and how computing might learn from neuroscience.
When we think about big data, we usually think about the web: the billions of users of social media, the sensors on millions of mobile phones, the thousands of contributions to Wikipedia, and so forth. Due to recent innovations, web-scale data can now also come from a camera pointed at a small, but extremely complex object: the brain. New progress in distributed computing is changing how neuroscientists work with the resulting data — and may, in the process, change how we think about computation. Read more…