Securing the cloud

Homomorphic encryption is one of the most exciting new research topics in cryptography, which promises to make perfectly secure. With it, a Web user would send encrypted data to a server in the cloud, which would process it without decrypting it and send back a still-encrypted result. Sometimes, however, the server needs to know something about the data it’s handling. Otherwise, some computational tasks become prohibitively time consuming — if not outright impossible. Suppose, for instance, that the task you’ve outsourced to the cloud is to search a huge encrypted database for the handful of records that match an encrypted search term. Homomorphic encryption ensures that the server has no idea what the search term is or which records match it. As a consequence, however, it has no choice but to send back information on every record in the database. The user’s computer can decrypt that information to see which records matched and which didn’t, but then it’s assuming much of the computational burden that it was trying to offload to the cloud in the first place.Last week, at the Association for Computing Machinery’s 45th Symposium on the Theory of Computing — the premier conference in theoretical computer science — researchers from MIT’s Computer Science and Artificial Intelligence Laboratory, together with colleagues at the University of Toronto and Microsoft Research, presented a new encryption scheme that solves this problem. Known as a functional-encryption scheme, it allows the cloud server to run a single, specified computation on the homomorphically encrypted result — asking, say, “Is this record a match?” or “Is this email spam?” — without being able to extract any other information about it.“This is a very, very general paradigm,” says Shafi Goldwasser, the RSA Professor of Electrical Engineering and Computer Science, one of the paper’s co-authors and, together with her fellow MIT professor Silvio Micali, the most recent recipient of the Turing Award, the highest award in computer science. “Say we’re talking about the surveillance cameras of the future, which come up with encrypted images. Why would we want to do that? It’s a question of liberty versus safety. If you’re looking for a suspect, you might be interested in doing some computations on an encrypted image, to match to the subject. Another possibility would be a medical database, where all the information is encrypted and … someone [runs] a drug study on those blood samples — but just that drug study, nothing else. Our result is in some sense the first result showing that you can do this very generally.”Joining Goldwasser on the paper are Raluca Ada Popa, a graduate student in the Department of Electrical Engineering and Computer Science, her advisor, associate professor Nickolai Zeldovich, and Yael Kalai of Microsoft Research and Vinod Vaikuntanathan of the University of Toronto, both of whom did their graduate work at MIT with Goldwasser.Near missesThe researchers built their functional-encryption scheme by fitting together several existing schemes, each of which has vital attributes of functional encryption, but none of which is entirely sufficient in itself. The first of those is homomorphic encryption.Another is what’s known as a garbled circuit, a technique developed in the mid-1980s and widely used in cryptography. A garbled circuit lets a user decrypt the result of one cryptographically protected operation on one cryptographically protected data item — say, “Is this record a match?” The problem is that, if the garbled circuit is used on a second data item — “How about this record?” — the security breaks.Moreover, a garbled circuit is a so-called private-key system, in which only the holder of a secret cryptographic key can encrypt data. Homomorphic encryption, by contrast, is intended as a public-key system — like most of the encryption schemes used to protect financial transactions on the Web. With public-key encryption, anyone can encrypt a message using a key that’s published online, but only the holder of the secret key can decrypt it.The final component technique is called attribute-based encryption. Attribute-based encryption is a public-key system, and it’s reusable. But unlike garbled circuits and homomorphic encryption, it can’t reveal the output of a function without revealing the input, too. The new system begins with homomorphic encryption and embeds the decryption algorithm in a garbled circuit. The key to the garbled circuit, in turn, is protected by attribute-based encryption. In some sense, the garbled circuit can, like all garbled circuits, be used only once. But the encryption schemes are layered in such a way that one use grants the server access to a general function rather than a single value. It can thus ask, of every record in a database, “Is this a match?”Zeldovich points out that since the scheme relies on homomorphic encryption, it shares the major drawback of existing homomorphic schemes: They’re still too computationally intensive to be practical. On the other hand, he says, “It’s so new, there are so many things that haven’t been explored — like, ‘How do you really implement this correctly?’ ‘What are the right mathematical constructions?’ ‘What are the right parameter settings?’” And, Popa adds, in the four years since the invention of the first fully homomorphic encryption scheme, “People have been shaving off many orders of magnitude in performance improvements.”Besides, even a currently impractical functional-encryption scheme is still a breakthrough. “Before, we didn’t even know if this was possible,” Popa says.Ran Canetti, a professor of computer science at Boston University, corroborates that assessment. “It’s an extremely surprising result,” he says. “I myself worked on this problem for a while, and I had no idea how to do it. So I was wowed. And it really opens up the door to many other applications.”One of those applications, Canetti says, is what’s known as program obfuscation, or disguising the operational details of a computer program so that it can’t be reverse-engineered. “Not obfuscating the way that people are doing it now, which is just scrambling up programs and hoping nobody will understand, and eventually, these are broken,” Canetti says, “but really obfuscating so that it’s cryptographically secure.”Canetti acknowledges that the researchers’ scheme won’t be deployed tomorrow. But “I’m sure it’s going to lead to more stuff,” he says. “It’s an enabler, and people will be building on it.”
Quelle: Massachusetts Institute of Technology

Detecting program-tampering in the cloud

For small and midsize organizations, the outsourcing of demanding computational tasks to the cloud — huge banks of computers accessible over the Internet — can be much more cost-effective than buying their own hardware. But it also poses a security risk: A malicious hacker could rent space on a cloud server and use it to launch programs that hijack legitimate applications, interfering with their execution.In August, at the International Cryptology Conference, researchers from MIT and Israel’s Technion and Tel Aviv University presented a new system that can quickly verify that a program running on the cloud is executing properly. That amounts to a guarantee that no malicious code is interfering with the program’s execution.The same system also protects the data used by applications running in the cloud, cryptographically ensuring that the user won’t learn anything other than the immediate results of the requested computation. If, for instance, hospitals were pooling medical data in a huge database hosted on the cloud, researchers could look for patterns in the data without compromising patient privacy.Although the paper reports new theoretical results (view PDF), the researchers have also built working code that implements their system. At present, it works only with programs written in the C programming language, but adapting it to other languages should be straightforward.The new work, like much current research on secure computation, requires that computer programs be represented as circuits. So the researchers’ system includes a “circuit generator” that automatically converts C code to circuit diagrams. The circuits it produces, however, are much smaller than those produced by its predecessors, so by itself, the circuit generator may find other applications in cryptography.Zero knowledgeAlessandro Chiesa, a graduate student in electrical engineering and computer science at MIT and one of the paper’s authors, says that because the new system protects both the integrity of programs running in the cloud and the data they use, it’s a good complement to the cryptographic technique known as homomorphic encryption, which protects the data transmitted by the users of cloud applications. On the paper, Chiesa joins Madars Virza, also a graduate student in electrical engineering and computer science; the Technion’s Daniel Genkin and Eli Ben-Sasson, who was a visiting scientist at MIT for the past year; and Tel Aviv University’s Eran Tromer. Ben-Sasson and Tromer were co-PIs on the project. The researchers’ system implements a so-called zero-knowledge proof, a type of mathematical game invented by MIT professors Shafi Goldwasser and Silvio Micali and their colleague Charles Rackoff of the University of Toronto. In its cryptographic application, a zero-knowledge proof enables one of the game’s players to prove to the other that he or she knows a secret key without actually divulging it.But as its name implies, a zero-knowledge proof is a more general method for proving mathematical theorems — and the correct execution of a computer program can be redescribed as a theorem. So zero-knowledge proofs are by definition able to establish whether or not a computer program is executing correctly.The problem is that existing implementations of zero-knowledge proofs — except in cases where they’ve been tailored to particular algorithms — take as long to execute as the programs they’re trying to verify. That’s fine for password verification, but not for a computation substantial enough that it might be farmed out to the cloud.The researchers’ innovation is a practical, succinct zero-knowledge proof for arbitrary programs. Indeed, it’s so succinct that it can typically fit in a single data packet.Linear thinkingAs Chiesa explains, his and his colleagues’ approach depends on a variation of what’s known as a “probabilistically checkable proof,” or PCP. “With a standard mathematical proof, if you want to verify it, you have to go line by line from the start to the end,” Chiesa says. “If you were to skip one line, potentially, that could fool you. Traditional proofs are very fragile in this respect.” “The PCP theorem says that there is a way to rewrite proofs so that instead of reading them line by line,” Chiesa adds, “what you can do is flip a few coins and probabilistically sample three or four lines and have a probabilistic guarantee that it’s correct.”The problem, Virza says, is that “the current known constructions of the PCP theorem, though great in theory, have quite bad practical realizations.” That’s because the theory assumes that an adversary who’s trying to produce a fraudulent proof has unbounded computational capacity. What Chiesa, Virza and their colleagues do instead is assume that the adversary is capable only of performing simple linear operations.“This assumption is, of course, false in practice,” Virza says. “So we use a cryptographic encoding to force the adversary to only linear evaluations. There is a way to encode numbers into such a form that you can add those numbers, but you can’t do anything else. This is how we sidestep the inefficiencies of the PCP theorem.”“I think it’s a breakthrough,” says Ran Canetti, a professor of computer science at Boston University who was not involved with the research. When the PCP theorem was first proved, Canetti says, “nobody ever thought that this would be something that would be remotely practical. They’ve become a little bit better over the years, but not that much better.”“Four or five years ago,” Canetti adds, “these guys wrote on the flag the crazy goal of trying to make [proofs for arbitrary programs] practical, and I must say, I thought, ‘They’re nuts.’ But they did it. They actually have something that works.”
Quelle: Massachusetts Institute of Technology

Protecting data in the cloud

— outsourcing computational tasks over the Internet — could give home-computer users unprecedented processing power and let small companies launch sophisticated Web services without building massive server farms.But it also raises privacy concerns. A bank of cloud servers could be running applications for 1,000 customers at once; unbeknownst to the hosting service, one of those applications might have no purpose other than spying on the other 999.Encryption could make cloud servers more secure. Only when the data is actually being processed would it be decrypted; the results of any computations would be re-encrypted before they’re sent off-chip.In the last 10 years or so, however, it’s become clear that even when a computer is handling encrypted data, its memory-access patterns — the frequency with which it stores and accesses data at different memory addresses — can betray a shocking amount of private information. At the International Symposium on Computer Architecture in June, MIT researchers described a new type of secure hardware component, dubbed Ascend, that would disguise a server’s memory-access patterns, making it impossible for an attacker to infer anything about the data being stored. Ascend also thwarts another type of attack, known as a timing attack, which attempts to infer information from the amount of time that computations take.Computational trade-offSimilar designs have been proposed in the past, but they’ve generally traded too much computational overhead for security. “This is the first time that any hardware design has been proposed — it hasn’t been built yet — that would give you this level of security while only having about a factor of three or four overhead in performance,” says Srini Devadas, the Edwin Sibley Webster Professor of Electrical Engineering and Computer Science, whose group developed the new system. “People would have thought it would be a factor of 100.”The “trivial way” of obscuring memory-access patterns, Devadas explains, would be to request data from every address in the memory — whether a memory chip or a hard drive — and throw out everything except the data stored at the one address of interest. But that would be much too time-consuming to be practical.What Devadas and his collaborators — graduate students Ling Ren, Xiangyao Yu and Christopher Fletcher, and research scientist Marten van Dijk — do instead is to arrange memory addresses in a data structure known as a “tree.” A family tree is a familiar example of a tree, in which each “node” (in this example, a person’s name) is attached to only one node above it (the node representing the person’s parents) but may connect to several nodes below it (the person’s children).With Ascend, addresses are assigned to nodes randomly. Every node lies along some “path,” or route through the tree, that starts at the top and passes from node to node, without backtracking, until arriving at a node with no further connections. When the processor requires data from a particular address, it sends requests to all the addresses in a path that includes the one it’s really after.To prevent an attacker from inferring anything from sequences of memory access, every time Ascend accesses a particular memory address, it randomly swaps that address with one stored somewhere else in the tree. As a consequence, accessing a single address multiple times will very rarely require traversing the same path.Less computation to disguise an addressBy confining its dummy requests to a single path, rather than sending them to every address in memory, Ascend exponentially reduces the amount of computation required to disguise an address. In a separate paper, which is as-yet unpublished but has been posted online, the researchers prove that querying paths provides just as much security as querying every address in memory would.Ascend also protects against timing attacks. Suppose that the computation being outsourced to the cloud is the mammoth task of comparing a surveillance photo of a criminal suspect to random photos on the Web. The surveillance photo itself would be encrypted, and thus secure from prying eyes. But spyware in the cloud could still deduce what public photos it was being compared to. And the time the comparisons take could indicate something about the source photos: Photos of obviously different people could be easy to rule out, but photos of very similar people might take longer to distinguish.So Ascend’s memory-access scheme has one final wrinkle: It sends requests to memory at regular intervals — even when the processor is busy and requires no new data. That way, attackers can’t tell how long any given computation is taking.
Quelle: Massachusetts Institute of Technology

Computing at full capacity

According to a 2014 study from NRDC and Anthesis, in 2013 U.S. data centers burned 91 billion kilowatt-hours of electricity, enough to power every household in New York City twice over. That figure is expected to rise to 140 billion by 2020. While improved energy efficiency practices could go a long way toward lowering this figure, the problem is greatly exacerbated by the underutilization of servers, including an estimated 30 percent of servers that are still plugged in, but are no longer performing any services, the study says.

In another 2014 study, tech research firm Gartner, Inc., found that data center systems collectively represent a $143 billion market. With enterprise software adding $320 billion to that and IT services another $963 billion, the overall IT industry represents a whopping $3.8 trillion market.

Companies are increasingly seeking new ways to cut costs and extract the largest possible value from their IT infrastructure. Strategies include placing data centers in cooler climates, switching to more affordable open source software, and virtualizing resources to increase utilization. These solutions just scratch the surface, however.

An MIT-connected startup called Jisto offers businesses a new tool for cutting data center and cloud costs while improving resource utilization. Jisto manages existing enterprise applications by automatically wrapping them in Jisto-managed Docker containers, and intelligently deploying them across all available resources using automated real-time deployment, monitoring, and analytics algorithms. As the resource utilization profile changes for each server or different parts of the network and storage, Jisto elastically scales its utilization in real-time to compensate.

“We’re helping organizations get higher utilization of their data center and cloud resources without worrying about resource contention,” says Jisto CEO and co-founder Aleksandr (Sasha) Biberman. So far, the response has been promising. Jisto was a Silver Winner in the 2014 MassChallenge, and early customers include data-intensive companies such as banks, pharmaceutical companies, biotech firms, and research institutions.

“There’s pressure on IT departments from two sides: How can they more efficiently reduce data center expenditures, and how can they improve productivity by giving people better access to resources,” Biberman says. “In some cases, Jisto can double the productivity with the same resources just by making better use of idle capacity.”

Biberman praises the MIT Industrial Liaison Program and Venture Mentoring Service for hosting networking events and providing connections. “The ILP gave us connections to companies that we would have never otherwise have connected to all around the world,” he says. “It turned us into a global company.”

Putting idle servers back to work

The idea for Jisto came to Biberman while he was a postdoc in electrical engineering at MIT Research Lab of Electronics (RLE), studying silicon photonic communications. While researching how optical technology could improve data center performance and efficiency, he discovered an even larger problem: underutilization of server resources.

“Even with virtualization, companies use only 20 to 50 percent of in-house server capacity,” Biberman says. “Collectively, companies are wasting more than $100 billion annually on unused cycles. The public cloud is even worse, where utilization runs at 10 to 40 percent.”

In addition to the problem of sheer waste, Biberman also discovered that workload resources are often poorly managed. Even when more than a half of a company’s resources are sitting idle, workers often complain they can’t get enough access to servers when they need them.

Around the time of Biberman’s realization, he and his long-time friend Andrey Turovsky, a Cornell University-educated tech entrepreneur, and now Jisto CTO and co-founder, had been brainstorming some startup ideas. They had just developed a lightweight platform to automatically deploy and manage applications using virtual containers, and they decided to apply it to the utilization and workload management problem.

Underutilization of resources is less a technical issue, than a “corporate risk aversion strategy,” Biberman says. Companies tend to err on the side of caution when deploying resources and typically acquire many more servers than they need.

“We started seeing some crazy numbers in data center and cloud provisioning,” Biberman explains. “Typically, companies provision for twice as much as they need. One company looks at last year’s peak loads, and overprovisions above that by a factor of four for the next year. Companies always plan for a worst-case scenario spike. Nobody wants to be the person who hasn’t provisioned enough resources, so critical applications can’t run. Nobody gets fired for overprovisioning.”

Despite overprovisioning, users in most of the same organizations complain about lack of access to computing resources, says Biberman: “When you ask companies if they have enough resources to run applications, they typically say they want more even though their resources are sitting there going to waste.”

This paradox emerges from the common practice of splitting access into different resource groups, which have different levels of access to various cluster nodes. “It’s tough to fit your work into your slice of the pie,” Biberman says. “Say my resource group has access to five servers, and it’s agreed that I use them on Monday, and someone else takes Tuesday, and so on. But if I can’t get to my project on Monday, those servers are sitting completely idle, and I may have to wait a week. Maybe the person using it on Tuesday only needs one of the five servers, so four will sit idle, and maybe the guy using it the next day realizes he really needs 10 or 20 servers, not just the five he’s limited to.”

Jisto breaks down the artificial static walls created with ownership profiles and replaces them with a more dynamic environment. “You can still have priority during your server time, but if you don’t use it, someone else can,” Biberman explains. “That means people can sometimes get access to more servers than were allotted. If there’s a mission-critical application that generates a spike we can’t predict, we have an elastic method to quickly back off and give it priority.”

Financial services companies are using Jisto to free up compute cycles for Monte Carlo simulations that could benefit from many more servers and nodes. Pharma and life science companies, meanwhile, use a similar strategy to do faster DNA sequencing. “The more nodes you have, the more accurately you can run a simulation,” Biberman says. “That’s a huge advantage.”

Docker containers for the enterprise

Jisto is not the only cloud-computing platform that claims to improve resource utilization and reduce costs. The problem with most, however, is that “if you have a really quick spike in workload, there’s not enough time to make intelligent decisions about what to do,” Biberman says. “With Jisto, an automatic real-time decision-making process kicks in, enabling true elasticity across the entire data center with granularity as fine as a single core of a CPU.”

Jisto not only monitors CPU usage but other parameters such as memory, network bandwidth, and storage. “If there’s an important memory transfer happening that requires a lot of bandwidth, Jisto backs off, even if there’s plenty of CPU power available,” Biberman says. “Jisto can make intelligent decisions about where to send jobs based on all these dynamic factors. As soon as something changes, Jisto decides whether to stop the workload, pause it, or reduce resources. Do you transfer it to another server? Do you add redundancy to reduce the latency tail? People don’t have to make and implement those decisions.”

The platform also integrates rigorous security provisions, says Biberman. IT directors are understandably cautious about bringing third-party software into their complex data center ecosystems, which are often protected by firewall and regulation settings. Jisto, however, can quickly prove with a beta test how the software can spin its magic without interfering with mission-critical resources, he adds.

Jisto’s unobtrusiveness is largely due to its use of Docker containers. “Docker has nice APIs and makes the process much easier, both for us as developers and for Jisto customers,” Biberman explains. “Docker is very portable — if you can run it on Linux, you can run it on Docker — and it doesn’t care if you’re running it on a local data center, a private cloud, or on Amazon. With containers, we don’t need to do something complicated like run a VM inside another VM. Docker gives us a lightweight way to let people use the environment that’s already set up.”

Based in Cambridge, Massachusetts, Jisto was the first, and remains one of few, Docker-based startups in this region.

Moving up to the cloud

Companies are increasingly saving on data center costs by using public cloud resources in a hybrid strategy during peak demand. Jisto can help bridge the gap with better efficiency and flexibility, says Biberman. “If you’re a bank, you might have too many regulations on your data to use the public cloud, but most companies can gain efficiencies with public clouds while still keeping their private cloud for confidential, regulated, or mission-critical tasks.”

Jisto operates essentially the same whether it’s running on-premises, or in a private, public, or hybrid cloud. Companies that exceed the peak level of their private data center can now “burst out” onto the public cloud and take advantage of the elastic nature of services such as Amazon, says Biberman. “Some companies provision hundreds of thousands of nodes on Amazon,” he adds. The problem is that Amazon charges by the hour. “If a company only needs five minutes of processing, as many as 100,000 nodes would sit idle for 55 minutes.”

Jisto has recently begun to talk to companies that do cloud infrastructure as a service, explaining how Jisto can reprovision wasted resources and let someone else use them. According to Biberman, it’s only a matter of time before competitive pressures lead a cloud provider to use something like Jisto.

MIT Startup Exchange (STEX) is an initiative of MIT’s Industrial Liaison Program (ILP) that seeks to connect ILP member companies with MIT-connected startups. Visit the STEX website and log in to learn more about Jisto and other startups on STEX.
Quelle: Massachusetts Institute of Technology

Diagnosing “broken" buildings to make them greener

The co-founders of MIT spinout KGS Buildings have a saying: “All buildings are broken.” Energy wasted through faulty or inefficient equipment, they say, can lead to hundreds of thousands of dollars in avoidable annual costs.

That’s why KGS aims to “make buildings better” with cloud-based software, called Clockworks, that collects existing data on a building’s equipment — specifically in HVAC (heating, ventilation, and air conditioning) equipment — to detect leaks, breaks, and general inefficiencies, as well as energy-saving opportunities.

The software then translates the data into graphs, metrics, and text that explain monetary losses, where it’s available for building managers, equipment manufacturers, and others through the cloud.

Building operators can use that information to fix equipment, prioritize repairs, and take efficiency measures — such as using chilly outdoor air, instead of air conditioning, to cool rooms.

“The idea is to make buildings better, by helping people save time, energy, and money, while providing more comfort, enjoyment, and productivity,” says Nicholas Gayeski SM ’07, PhD ’10, who co-founded KGS with Sian Kleindienst SM ’06, PhD ’10 and Stephen Samouhos ’04, SM ’07, PhD ’10.

The software is now operating in more than 300 buildings across nine countries, collecting more than 2 billion data points monthly. The company estimates these buildings will save an average of 7 to 9 percent in avoidable costs per year; the exact figure depends entirely on the building. 

“If it’s a relatively well-performing building already, it may see lower savings; if it’s a poor-performing building, it could be much higher, maybe 15 to 20 percent,” says Gayeski, who graduated from MIT’s Building Technology Program, along with his two co-founders.

Last month, MIT commissioned the software for more than 60 of its own buildings, monitoring more than 7,000 pieces of equipment over 10 million square feet. Previously, in a year-long trial for one MIT building, the software saved MIT $286,000.  

Benefits, however, extend beyond financial savings, Gayeski says. “There are people in those buildings: What’s their quality of life? There are people who work on those buildings. We can provide them with better information to do their jobs,” he says.

The software can also help buildings earn additional incentives by participating in utility programs. “We have major opportunities in some utility territories, where energy-efficiency has been incentivized. We can help buildings meet energy-efficiency goals that are significant in many states, including Massachusetts,” says Alex Grace, director of business development for KGS.

Other customers include universities, health-care and life-science facilities, schools, and retail buildings.

Equipment-level detection

Fault-detection and diagnostics research spans about 50 years — with contributions by early KGS advisors and MIT professors of architecture Les Norford and Leon Glicksman — and about a dozen companies now operate in the field.

But KGS, Gayeski says, is one of a few ventures gathering “equipment-level data,” gathered through various sensors, actuators, and meters attached to equipment that measure functionality.

Clockworks sifts through that massive store of data, measuring temperatures, pressures, flows, set points, and control commands, among other things. It’s able to gather a few thousand data points every five minutes — which is a finer level of granularity than meter-level analytics software that may extract, say, a data point every 15 minutes from a utility meter.

“That gives a lot more detail, a lot more granular information about how things are operating and could be operating better,” Gayeski says. For example, Clockworks may detect specific leaky valves or stuck dampers on air handlers in HVAC units that cause excessive heating or cooling.

To make its analyses accurate, KGS employs what Gayeski calls “mass customization of code.” The company has code libraries for each type of equipment it works with — such as air handlers, chillers, and boilers — that can be tailored to specific equipment that varies greatly from building to building.

This makes Clockworks easily scalable, Gayeski says. But it also helps the software produce rapid, intelligent analytics — such as accurate graphs, metrics, and text that spell out problems clearly.

Moreover, it helps the software to rapidly equate data with monetary losses. “When we identify that there’s a fault with the right data, we can tell people right away this is worth, say, $50 a day or this is worth $1,000 a day — and we’ve seen $1,000-a-day faults — so that allows facilities managers to prioritize which problems get their attention,” he says.

KGS Buildings’ foundation

The KGS co-founders met as participants in the MIT entry for the 2007 Solar Decathlon — an annual competition where college teams build small-scale, solar-powered homes to display at the National Mall in Washington. Kleindienst worked on lighting systems, while Samouhos and Gayeski worked on mechanical design and energy-modeling.

After the competition, the co-founders started a company with a broad goal of making buildings better through energy savings. While pursuing their PhDs, they toyed with various ideas, such as developing low-cost sensing technology with wireless communication that could be retrofitted on to older equipment.

Seeing building data as an emerging tool for fault-detection and diagnostics, however, they turned to Samouhos’ PhD dissertation, which focused on building condition monitoring. It came complete with the initial diagnostics codes and a framework for an early KGS module.

“We all came together anticipating that the building industry was about to change a lot in the way it uses data, where you take the data, you figure out what’s not working well, and do something about it,” Gayeski says. “At that point, we knew it was ripe to move forward.”

Throughout 2010, they began trialing software at several locations, including MIT. They found guidance among the seasoned entrepreneurs at MIT’s Venture Mentoring Service — learning to fail fast, and often. “That means keep at it, keep adapting and adjusting, and if you get it wrong, you just fix it and try again,” Gayeski says.

Today, the company — headquartered in Somerville, Mass., with 16 employees — is focusing on expanding its customer base and advancing its software into other applications. About 180 new buildings were added to Clockworks in the past year; by the end of 2014, KGS projects it could deploy its software to 800 buildings. 

“Larger companies are starting to catch on,” Gayeski says. “Major health-care institutions, global pharmaceuticals, universities, and [others] are starting to see the value and deciding to take action — and we’re starting to take off.”

Liberating data

By bringing all this data about building equipment to the cloud, the technology has plugged into the “Internet of things” — a concept where objects would be connected, via embedded chips and other methods, to the Internet for inventory and other purposes.

Data on HVAC systems have been connected through building automation for some time. KGS, however, can connect that data to cloud-based analytics and extract “really rich information” about equipment, Gayeski says. For instance, he says, the startup has quick-response codes — like a barcode — for each piece of equipment it measures, so people can read all data associated with it.

“As more and more devices are readily connected to the Internet, we may be tapping straight into those, too,” Gayeski says. 

“And that data can be liberated from its local environment to the cloud,” Grace adds. 

Down the road, as technology to monitor houses — such as automated thermostats and other sensors — begins to “unlock the data in the residential scale,” Gayeski says, “KGS could adapt over time into that space, as well.”
Quelle: Massachusetts Institute of Technology

How can we protect our information in the era of cloud computing?

In an article published in the Proceedings of the Royal Society A, Professor Jon Crowcroft argues that by parcelling and spreading data across multiple sites, and weaving it together like a tapestry, not only would our information be safer, it would be quicker to access, and could potentially be stored at lower overall cost.
The internet is a vast, decentralised communications system, with minimal administrative or governmental oversight. However, we increasingly access our information through cloud-based services, such as Google Drive, iCloud and Dropbox, which are very large centralised storage and processing systems. Cloud-based services offer convenience to the user, as their data can be accessed from anywhere with an internet connection, but their centralised nature can make them vulnerable to attack, such as when personal photos of mostly young and female celebrities were leaked last summer after their iCloud accounts were hacked.
Storing information in the cloud makes it easily accessible to users, while removing the burden of managing it; and the cloud’s highly centralised nature keeps costs low for the companies providing the storage. However, centralised systems can lack resilience, meaning that service can be lost when any one part of the network access path fails.
Centralised systems also give a specific point to attack for those who may want to access them illegally. Even if data is copied many times, if all the copies have the same flaw, they are all vulnerable. Just as a small gene pool places a population at risk from a change in the environment, such as a disease, the lack of variety in centralised storage systems places information at greater risk of theft.
The alternative is a decentralised system, also known as a peer-to-peer system, where resources from many potential locations in the network are mixed, rather than putting all one’s eggs in one basket.
The strength of a peer-to-peer system is that its value grows as the number of users increases: all producers are also potential consumers, so each added node gives the new producer as many customers as are already on the network.
“Since all the members of a peer-to-peer network are giving as well as consuming resources, it quickly overtakes a centralised network in terms of its strength,” said Crowcroft, of the University’s Computer Laboratory.
The higher reliability and performance of fibre to the home, the availability of 4G networks, and IPv6 (Internet Protocol version 6) are all helping to make decentralised networks viable. In practice, a user would carry most of the data they need to access immediately with them on their mobile device, with their home computer acting as the ‘master’ point of contact.
“Essentially, data is encoded redundantly, but rather than making many copies, we weave a tapestry using the bits that represent data, so that threads making up particular pieces of information are repeated but meshed together with threads making up different pieces of information,” said Crowcroft. “Then to dis-entangle a particular piece of information, we need to unpick several threads.”
Varying the ways that our information is stored or distributed is normally done to protect against faults in the network, but it can also improve the privacy of our data. In a decentralised system where data is partitioned across several sites, any attacker attempting to access that data has a much more complex target – the attacker has to know where all bits of the information are, as opposed to using brute force at one point to access everything. “The more diversity we use in a peer-to-peer system, the closer we get to an ideal in terms of resilience and privacy,” said Crowcroft.
A peer-to-peer system could also be built at a lower overall cost than a centralised system, argues Crowcroft, since no ‘cache’ is needed in order to store data near the user. To the end user, costs could be as low as a pound per month, or even free, much lower than monthly internet access costs or mobile tariffs.
“We haven’t seen massive take-up of decentralised networks yet, but perhaps that’s just premature,” said Crowcroft. “We’ve only had these massive centralised systems for about a decade, and like many other utilities, the internet will most likely move away from centralisation and towards decentralisation over time, especially as developments in technology make these systems attractive for customers.”
Private information would be much more secure if individuals moved away from cloud-based storage towards peer-to-peer systems, where data is stored in a variety of ways and across a variety of sites, argues a University of Cambridge researcher.
The more diversity we use in a peer-to-peer system, the closer we get to an ideal in terms of resilience and privacyJon Crowcroftg4ll4isPrivacyThe text in this work is licensed under a Creative Commons Licence. If you use this content on your site please link back to this page. For image rights, please see the credits associated with each individual image.

YesLicense type: Attribution-ShareAlike
Quelle: University of Cambridge

Trust on the wild web

Mark Zuckerberg is the world’s youngest billionaire. He got there by founding facebook.com, one of the biggest beasts in the Internet jungle. In the early days, so the story goes, he boasted to a friend on instant messenger that he had the personal details of over 4,000 students in Harvard, and if he ever wanted to know anything he should get in touch. Understandably, his incredulous friend wanted to know how Zuckerberg had access to this information. His reply? ‘People just submitted it. I don’t know why. They trust me, dumbf***s.’
The online environment is no longer merely an aid to living well offline; for many, it has become a forum where much of life is now conducted. But one issue that raises its head again and again is this question of trust on the Internet.
Examining whether and how we can design the Internet for online trust is the focus of my research in the Faculty of Philosophy, supervised by Dr Alex Oliver. The project is sponsored by Microsoft Research, whose Socio-Digital Systems group in Cambridge looks at how technology interacts with human values.
The research is a chance to do some practical philosophy, reflecting on and engaging with applied issues. And as the Internet increasingly becomes a more pervasive part of our lives, issues of trust online are only going to grow in importance. So there is a unique and timely opportunity – and challenge – to break new terrain.

Trusting me, trusting you
It is easily overlooked, but when you stop to think, it is striking how much we trust to other people. It is a fundamental precondition for the smooth functioning of society. Like the air we breathe, or the cement in brickwork, trust is both essential and usually taken for granted.
One consequence is that we tend to notice our reliance on trust only when things go wrong. And although it is easy to eulogise trust, it is not always appropriate. Trusting the untrustworthy is often a dramatically bad idea. But distrusting the trustworthy may have equally serious consequences.
Certainly, most people want to live in a world where it makes sense to trust people, and for people to trust them. But they also don’t want to be taken for a ride. So we have to work out when trust is appropriate.
The trouble is, it is much harder to work out online when trust is appropriate and when not. It is much more difficult to determine online whether a particular person is trustworthy – much of the personal and social context of offline forms of interaction are stripped away in cyberspace, and online identities can be less stable.
But perhaps more seriously, it is still relatively unclear what the norms and mores are that govern appropriate behaviour online. This applies both to the informal norms that spontaneously arise in interpersonal interaction, and also to the apparatus of formal law.
The web, in this sense, is a bit like the Wild West. It is not that life is impossible there – far from it. Indeed, it’s often pretty flamboyant and colourful, and a stimulating place to be. But people can also act unpredictably, and there is little recourse for those who get stung.

Building trust
One moral of the story about the Facebook founder’s comment is that you’ve got to be careful who you trust online. That’s obvious enough, and no different to what we tell our children.
But there are some more challenging issues. For the online world has an important feature: it is malleable. How something is built often serves particular ends, whether intended or not, and these ends in turn serve to realise particular visions of how people ought to live. Were I a metalsmith, for instance, I would rather make ploughs than thumbscrews – I don’t want to contribute to making a world where thumbscrews are plentiful.
This applies to contemporary technologies too. At the last count, 500 million people now have their social relationships partially structured by Zuckerberg’s vision of connecting people, according to whether they have confirmed or ignored the one-size-fits-all ‘friends request’ on facebook.com. The basic IP/TCP structures of the Internet were built according to a broadly libertarian vision widely shared among the early computer science pioneers, which denies central control or ownership in order to facilitate free expression.
So the more pertinent question is: can we build the Internet in a way that facilitates well-placed trust, and encourages trustworthiness? In short, can we design for online trust? To answer this, we need to look at why people are trustworthy and untrustworthy; what counts as good evidence for a person’s trustworthiness online; the effects of online anonymity and pseudonymity; and the role of institutions in grounding trustworthiness. For instance, one mechanism through which we can secure others’ trustworthiness is to develop better online reputation systems and track past conduct.
These questions cannot be answered once and for all. Technology is dynamic: , for instance, is considered by many to be a step change in the way we compute, and it too raises specific questions around trust (see panel). As technology changes, so too will the philosophical challenges. The hope is that collaborative work between computer engineers, lawyers and philosophers can help to make the Internet a safer place.

For further information, please contact Tom Simpson (tws21@cam.ac.uk), whose PhD research in the Faculty of Philosophy (www.phil.cam.ac.uk/) is being sponsored by Microsoft Research Cambridge. His article on ‘e-Trust and Reputation’ is published in Ethics and Information Technology.

Philosopher Tom Simpson asks: can we build a trustworthy and safe Internet?
Engineering is always about solving problems for people and the society in which they live. Philosophy can help understand what those problems are and how they are to be solved.

Professor Richard Harper, Microsoft Research Cambridge©iStockPhoto.com/Amanda RohdeDigital trustCloud computingCloud computing is widely heralded as one of the most radical changes to the way we compute, and its full impact is thought to be just around the corner. First and foremost, the cloud is a change in the geography of computing – instead of having your PC store your data and run everything, your computing will be done on banks of servers and accessed remotely. Along with the change in geography, the move to the cloud is also a change in the scale of computing, with access to far more powerful computing facilities than ever before.
But the cloud raises a host of philosophical issues, particularly questions of responsibility. Who should own what data? When are ‘crowd-sourcing’ techniques appropriate, and when not? What are the effects of more powerful techniques of profiling individuals? What happens to privacy when we compute in the cloud?
To discuss these and related issues, the Faculty of Philosophy and Microsoft Research are co-hosting an international conference in Cambridge, gathering together leading philosophers and practitioners. Two open lectures will be held on the evenings of 5 and 6 April 2011.
For further details, please visit trustandcloudcomputing.org.uk

This work is licensed under a Creative Commons Licence. If you use this content on your site please link back to this page.
Yes
Quelle: University of Cambridge

Privacy by design

Online services that store our personal information have proliferated, yet the technology to underpin how our privacy is safeguarded has lagged behind. This was the conclusion of a 2008 report by the UK’s Information Commissioner’s Office, a body set up to uphold privacy for individuals, which pressed for “the evolution of a new approach to the management of personal information that ingrains privacy principles into every part of every system in every organisation.”

This ethos underpins research led by Professor Jon Crowcroft, the Marconi Professor of Communications Systems in the Computer Laboratory. Two projects he leads aim to minimise privacy risks, and at the heart of both is the concept of ‘privacy by design’.

“Privacy by design means that it’s in-built as part of the technology, rather than bolted on in order to comply with data protection laws,” he explained. “With privacy by design, it would simply not be possible for incidents such as the leaking of LinkedIn passwords to happen.”

One project is tackling the challenge of how to maintain privacy when all your data are stored by a central service – the so-called cloud. Anyone who stores images on flickr, or accesses emails from a central server, is , and today many businesses are turning to centralised data centres as an economic means of storing their information. However, concerns have also been raised about the scale of control that cloud service providers wield over the data they store and can potentially monitor.

Crowcroft and colleague Dr Anil Madhavapeddy are building technologies to support the control of networked personal data as part of a five-year £12 million research hub (‘Horizon’), which is led by the University of Nottingham and funded by the Engineering and Physical Sciences Research Council (EPSRC). The research is driven by the overarching concept of a lifelong contextual footprint – the idea that each of us throughout our lifetime will lay down a digital trail that captures our patterns of interaction with digital services – and how best to protect this.

A second project, FRESNEL (for ‘Federated Secure Sensor Network Laboratory’), is focusing on privacy in networks that people use to modify their heating, lighting and home entertainment when they are not at home, as well as networks that monitor traffic flow and air quality, and enable a doctor in hospital to check a patient’s health at home.

“Current technologies have usually been devised for single-owner sensor networks that are deployed and managed by a central controlling entity, usually a company that has set themselves up to offer this capability,” he said. “They don’t have the right scalability and security required to deal with a secure multi-purpose federated sensor network, running different applications in parallel. Our aim is to build a network framework with multiple applications sharing the same resources.”

With funding from EPSRC, Crowcroft and Dr Cecilia Mascolo and colleagues, working with Dr Ian Brown at the University of Oxford and industrial project partners, now have a demonstrator program in operation that is currently being evaluated through a large-scale federation of sensor networks across the University of Cambridge.

The aim of these projects, explained Crowcroft, is not to lock up personal data, removing the ability to socialise it, but rather to support systems that process data without sacrificing privacy: “We are building technologies to support lifelong control of networked personal data. For instance, a significant driver behind social networking has been the ecosystem of data processors that aggregate and provide services such as recommendations, location searches or messaging. But the big drawback is that users have to divulge more of their personal data to a third party than is necessary, because of the difficulty of distinguishing what is needed. Our research starts from a single premise – that individuals require control over access to, and use of, their personal data for ever.”

Crowcroft and colleagues have launched a not-for-profit foundation, Digital Life Foundation, which will build an open-source community around these technologies.

For more information, please contact Louise Walsh (louise.walsh@admin.cam.ac.uk) at the University of Cambridge Office of External Affairs and Communications.
New research aims to ensure that we can exploit the full benefits of the digital world and still protect our online privacy.
We are building technologies to support lifelong control of networked personal data.Professor Jon Crowcroft©iStockPhoto.com/Marilyn NievesOnline privacy

This work is licensed under a Creative Commons Licence. If you use this content on your site please link back to this page.
Yes
Quelle: University of Cambridge

Cloud computing and the philosophy of trust

The event takes place today and tomorrow, and is hosted by the Faculty of Philosophy, University of Cambridge, and supported by Microsoft Research. An international workshop of world-renowned philosophers, ethicists, sociologists and practitioners will discuss the philosophical issues surrounding , a concept that has been described as one of the most radical changes to the way we compute. There will also be two public lectures at Corpus Christi, Cambridge.

Cloud computing is first and foremost a change in the geography of computing. Instead of the hardware on your computer doing the computing, the data storage and processing are carried out by hardware held in a different location. Facebook, Gmail and Flickr are well-known examples of computing in the cloud; a widespread move to cloud computing would see third-party servers providing nearly all computing needs, with users accessing software and data as needed.

Benefits claimed for cloud computing include access to far more powerful computing facilities than ever before, convenience and reliability of communications, greater flexibility for a mobile workforce, and a cost-effective alternative for businesses needing to maintain an up-to-date IT infrastructure.

But the provision of computing as a utility also raises philosophical issues, particularly questions of responsibility. Among these are who should own what data, and what happens to privacy when we compute in the cloud? How do we ensure the trustworthiness of those who manage the cloud, so that people use it confidently? And what is it about their computing practices that lead people to want the cloud?

To discuss these and related issues, the conference has gathered together delegates from institutions such as Massachusetts Institute of Technology (MIT), Rutgers, Institute Marcel Mauss, Paris, TU Delft, and the Universities of Cambridge and Oxford.

“We have a choice about how we build and regulate the cloud,” said Dr Alex Oliver, from Cambridge’s Faculty of Philosophy. “The aim of the event is to initiate a new discussion on how the internet and cloud computing is changing business and personal relationships in the cloud era.”

The public lecture this evening will be given by Dr David D. Clark, Senior Research Scientist, Computer Science and Artificial Intelligence Laboratory, MIT, at 6pm. Tomorrow’s lecture will be by Professor Ian Kerr, Canada Research Chair in Ethics, Law and Technology, University of Ottawa, at 5pm. Both lectures will be held in the McCrum Lecture Theatre, Corpus Christi, Cambridge.
Some of the world’s finest minds in academic philosophy are debating the impact of the internet and cloud computing in Cambridge this week.
We have a choice about how we build and regulate the cloud.Dr Alex OliverFaculty of PhilosophyCloud computing

This work is licensed under a Creative Commons Licence. If you use this content on your site please link back to this page.
Yes
Quelle: University of Cambridge