Skip to main content

Home/ Advanced Concepts Team/ Group items tagged continual

Rss Feed Group items tagged

dejanpetkow

Light Table - a new IDE concept - 2 views

  •  
    "Despite the dramatic shift toward simplification in software interfaces, the world of development tools continues to shrink our workspace with feature after feature in every release. Even with all of these things at our disposal, we're stuck in a world of files and forced organization - why are we still looking all over the place for the things we need when we're coding?"
Guido de Croon

Will robots be smarter than humans by 2029? - 2 views

  •  
    Nice discussion about the singularity. Made me think of drinking coffee with Luis... It raises some issues such as the necessity of embodiment, etc.
  • ...9 more comments...
  •  
    "Kurzweilians"... LOL. Still not sold on embodiment, btw.
  •  
    The biggest problem with embodiment is that, since the passive walkers (with which it all started), it hasn't delivered anything really interesting...
  •  
    The problem with embodiment is that it's done wrong. Embodiment needs to be treated like big data. More sensors, more data, more processing. Just putting a computer in a robot with a camera and microphone is not embodiment.
  •  
    I like how he attacks Moore's Law. It always looks a bit naive to me if people start to (ab)use it to make their point. No strong opinion about embodiment.
  •  
    @Paul: How would embodiment be done RIGHT?
  •  
    Embodiment has some obvious advantages. For example, in the vision domain many hard problems become easy when you have a body with which you can take actions (like looking at an object you don't immediately recognize from a different angle) - a point already made by researchers such as Aloimonos.and Ballard in the end 80s / beginning 90s. However, embodiment goes further than gathering information and "mental" recognition. In this respect, the evolutionary robotics work by for example Beer is interesting, where an agent discriminates between diamonds and circles by avoiding one and catching the other, without there being a clear "moment" in which the recognition takes place. "Recognition" is a behavioral property there, for which embodiment is obviously important. With embodiment the effort for recognizing an object behaviorally can be divided between the brain and the body, resulting in less computation for the brain. Also the article "Behavioural Categorisation: Behaviour makes up for bad vision" is interesting in this respect. In the field of embodied cognitive science, some say that recognition is constituted by the activation of sensorimotor correlations. I wonder to which extent this is true, and if it is valid for extremely simple creatures to more advanced ones, but it is an interesting idea nonetheless. This being said, if "embodiment" implies having a physical body, then I would argue that it is not a necessary requirement for intelligence. "Situatedness", being able to take (virtual or real) "actions" that influence the "inputs", may be.
  •  
    @Paul While I completely agree about the "embodiment done wrong" (or at least "not exactly correct") part, what you say goes exactly against one of the major claims which are connected with the notion of embodiment (google for "representational bottleneck"). The fact is your brain does *not* have resources to deal with big data. The idea therefore is that it is the body what helps to deal with what to a computer scientist appears like "big data". Understanding how this happens is key. Whether it is the problem of scale or of actually understanding what happens should be quite conclusively shown by the outcomes of the Blue Brain project.
  •  
    Wouldn't one expect that to produce consciousness (even in a lower form) an approach resembling that of nature would be essential? All animals grow from a very simple initial state (just a few cells) and have only a very limited number of sensors AND processing units. This would allow for a fairly simple way to create simple neural networks and to start up stable neural excitation patterns. Over time as complexity of the body (sensors, processors, actuators) increases the system should be able to adapt in a continuous manner and increase its degree of self-awareness and consciousness. On the other hand, building a simulated brain that resembles (parts of) the human one in its final state seems to me like taking a person who is just dead and trying to restart the brain by means of electric shocks.
  •  
    Actually on a neuronal level all information gets processed. Not all of it makes it into "conscious" processing or attention. Whatever makes it into conscious processing is a highly reduced representation of the data you get. However that doesn't get lost. Basic, low processed data forms the basis of proprioception and reflexes. Every step you take is a macro command your brain issues to the intricate sensory-motor system that puts your legs in motion by actuating every muscle and correcting every step deviation from its desired trajectory using the complicated system of nerve endings and motor commands. Reflexes which were build over the years, as those massive amounts of data slowly get integrated into the nervous system and the the incipient parts of the brain. But without all those sensors scattered throughout the body, all the little inputs in massive amounts that slowly get filtered through, you would not be able to experience your body, and experience the world. Every concept that you conjure up from your mind is a sort of loose association of your sensorimotor input. How can a robot understand the concept of a strawberry if all it can perceive of it is its shape and color and maybe the sound that it makes as it gets squished? How can you understand the "abstract" notion of strawberry without the incredibly sensible tactile feel, without the act of ripping off the stem, without the motor action of taking it to our mouths, without its texture and taste? When we as humans summon the strawberry thought, all of these concepts and ideas converge (distributed throughout the neurons in our minds) to form this abstract concept formed out of all of these many many correlations. A robot with no touch, no taste, no delicate articulate motions, no "serious" way to interact with and perceive its environment, no massive flow of information from which to chose and and reduce, will never attain human level intelligence. That's point 1. Point 2 is that mere pattern recogn
  •  
    All information *that gets processed* gets processed but now we arrived at a tautology. The whole problem is ultimately nobody knows what gets processed (not to mention how). In fact an absolute statement "all information" gets processed is very easy to dismiss because the characteristics of our sensors are such that a lot of information is filtered out already at the input level (e.g. eyes). I'm not saying it's not a valid and even interesting assumption, but it's still just an assumption and the next step is to explore scientifically where it leads you. And until you show its superiority experimentally it's as good as all other alternative assumptions you can make. I only wanted to point out is that "more processing" is not exactly compatible with some of the fundamental assumptions of the embodiment. I recommend Wilson, 2002 as a crash course.
  •  
    These deal with different things in human intelligence. One is the depth of the intelligence (how much of the bigger picture can you see, how abstract can you form concept and ideas), another is the breadth of the intelligence (how well can you actually generalize, how encompassing those concepts are and what is the level of detail in which you perceive all the information you have) and another is the relevance of the information (this is where the embodiment comes in. What you do is to a purpose, tied into the environment and ultimately linked to survival). As far as I see it, these form the pillars of human intelligence, and of the intelligence of biological beings. They are quite contradictory to each other mainly due to physical constraints (such as for example energy usage, and training time). "More processing" is not exactly compatible with some aspects of embodiment, but it is important for human level intelligence. Embodiment is necessary for establishing an environmental context of actions, a constraint space if you will, failure of human minds (i.e. schizophrenia) is ultimately a failure of perceived embodiment. What we do know is that we perform a lot of compression and a lot of integration on a lot of data in an environmental coupling. Imo, take any of these parts out, and you cannot attain human+ intelligence. Vary the quantities and you'll obtain different manifestations of intelligence, from cockroach to cat to google to random quake bot. Increase them all beyond human levels and you're on your way towards the singularity.
Thijs Versloot

Bacterial growth in space - 0 views

  •  
    According to the article it seems that the growth of biofilms in space (MIR, ISS) is continuing to be a challenge for any space missions. Hardly any information is present on what role gravity plays in their growth and development. Maybe detection is also one important aspect of this indeed (Tom?)
Athanasia Nikolaou

Nature Paper: Rivers and streams release more CO2 than previously believed - 6 views

  •  
    Another underestimated source of CO2, are turbulent waters. "The stronger the turbulences at the water's surface, the more CO2 is released into the atmosphere. The combination of maps and data revealed that, while the CO2 emissions from lakes and reservoirs are lower than assumed, those from rivers and streams are three times as high as previously believed." Alltogether the emitted CO2 equates to roughly one-fifth of the emissions caused by humans. Yet more stuff to model...
  • ...10 more comments...
  •  
    This could also be a mechanism to counter human CO2 emission ... the more we emit, the less turbulent rivers and stream, the less CO2 is emitted there ... makes sense?
  •  
    I guess there is a natural equilibrium there. Once the climate warms up enough for all rivers and streams to evaporate they will not contribute CO2 anymore - which stops their contribution to global warming. So the problem is also the solution (as always).
  •  
    "The source of inland water CO2 is still not known with certainty and new studies are needed to research the mechanisms controlling CO2 evasion globally." It is another source of CO2 this one, and the turbulence in the rivers is independent of our emissions in CO2 and just facilitates the process of releasing CO2 waters. Dario, if I understood correct you have in mind a finite quantity of CO2 that the atmosphere can accomodate, and to my knowledge this does not happen, so I cannot find a relevant feedback there. Johannes, H2O is a powerful greenhouse gas :-)
  •  
    Nasia I think you did not get my point (a joke, really, that Johannes continued) .... by emitting more CO2 we warm up the planet thus drying up rivers and lakes which will, in turn emit less CO2 :) No finite quantity of CO2 in the atmosphere is needed to close this loop ... ... as for the H2O it could just go into non turbulent waters rather than staying into the atmosphere ...
  •  
    Really awkward joke explanation: I got the joke of Johannes, but maybe you did not get mine: by warming up the planet to get rid of the rivers and their problems, the water of the rivers will be accomodated in the atmosphere, therefore, the greenhouse gas of water.
  •  
    from my previous post: "... as for the H2O it could just go into non turbulent waters rather than staying into the atmosphere ..."
  •  
    I guess the emphasis is on "could"... ;-) Also, everybody knows that rain is cold - so more water in the atmosphere makes the climate colder.
  •  
    do you have the nature paper also? looks like very nice, meticulous typically german research lasting over 10 years with painstakingly many researchers from all over the world involved .... and while important the total is still only 20% of human emissions ... so a variation in it does not seem to change the overall picture
  •  
    here is the nature paper : http://www.nature.com/nature/journal/v503/n7476/full/nature12760.html I appreciate Johannes' and Dario's jokes, since climate is the common ground that all of us can have an opinion, taking honours from experiencing weather. But, the same as if I am trying to make jokes for material science, or A.I. I take a high risk of failing(!) :-S Water is a greenhouse gas, rain rather releases latent heat to the environment in order to be formed, Johannes, nice trolling effort ;-) Between this and the next jokes to come, I would stop to take a look here, provided you have 10 minutes: how/where rain forms http://www.scribd.com/doc/58033704/Tephigrams-for-Dummies
  •  
    omg
  •  
    Nasia, I thought about your statement carefully - and I cannot agree with you. Water is not a greenhouse gas. It is instead a liquid. Also, I can't believe you keep feeding the troll! :-P But on a more topical note: I think it is an over-simplification to call water a greenhouse gas - water is one of the most important mechanisms in the way Earth handles heat input from the sun. The latent heat that you mention actually cools Earth: solar energy that would otherwise heat Earth's surface is ABSORBED as latent heat by water which consequently evaporates - the same water condenses into rain drops at high altitudes and releases this stored heat. In effect the water cycle is a mechanism of heat transport from low altitude to high altitude where the chance of infrared radiation escaping into space is much higher due to the much thinner layer of atmosphere above (including the smaller abundance of greenhouse gasses). Also, as I know you are well aware, the cloud cover that results from water condensation in the troposphere dramatically increases albedo which has a cooling effect on climate. Furthermore the heat capacity of wet air ("humid heat") is much larger than that of dry air - so any advective heat transfer due to air currents is more efficient in wet air - transporting heat from warm areas to a natural heat sink e.g. polar regions. Of course there are also climate heating effects of water like the absorption of IR radiation. But I stand by my statement (as defended in the above) that rain cools the atmosphere. Oh and also some nice reading material on the complexities related to climate feedback due to sea surface temperature: http://journals.ametsoc.org/doi/abs/10.1175/1520-0442(1993)006%3C2049%3ALSEOTR%3E2.0.CO%3B2
  •  
    I enjoy trolling conversations when there is a gain for both sides at the end :-) . I had to check upon some of the facts in order to explain my self properly. The IPCC report states the greenhouse gases here, and water vapour is included: http://www.ipcc.ch/publications_and_data/ar4/wg1/en/faq-2-1.html Honestly, I read only the abstract of the article you posted, which is a very interesting hypothesis on the mechanism of regulating sea surface temperature, but it is very localized to the tropics (vivid convection, storms) a region of which I have very little expertise, and is difficult to study because it has non-hydrostatic dynamics. The only thing I can comment there is that the authors define constant relative humidity for the bottom layer, supplied by the oceanic surface, which limits the implementation of the concept on other earth regions. Also, we may confuse during the conversation the greenhouse gas with the Radiative Forcing of each greenhouse gas: I see your point of the latent heat trapped in the water vapour, and I agree, but the effect of the water is that it traps even as latent heat an amount of LR that would otherwise escape back to space. That is the greenhouse gas identity and an image to see the absorption bands in the atmosphere and how important the water is, without vain authority-based arguments that miss the explanation in the end: http://www.google.nl/imgres?imgurl=http://www.solarchords.com/uploaded/82/87-33833-450015_44absorbspec.gif&imgrefurl=http://www.solarchords.com/agw-science/4/greenhouse--1-radiation/33784/&h=468&w=458&sz=28&tbnid=x2NtfKh5OPM7lM:&tbnh=98&tbnw=96&zoom=1&usg=__KldteWbV19nVPbbsC4jsOgzCK6E=&docid=cMRZ9f22jbtYPM&sa=X&ei=SwynUq2TMqiS0QXVq4C4Aw&ved=0CDkQ9QEwAw
Thijs Versloot

Forget 3D printing, here is solar printing - 3 views

  •  
    The work is based on the use of a "continuous flow" microreactor to produce nanoparticle inks that make solar cells by printing. In this process, simulated sunlight is focused on the solar microreactor to rapidly heat it, while allowing precise control of temperature to aid the quality of the finished product. The light in these experiments was produced artificially, but the process could be done with direct sunlight, and at a fraction of the cost of current approaches.
  •  
    anybody interested in having a closer look at this?
tvinko

Massively collaborative mathematics : Article : Nature - 28 views

  •  
    peer-to-peer theorem-proving
  • ...14 more comments...
  •  
    Or: mathematicians catch up with open-source software developers :)
  •  
    "Similar open-source techniques could be applied in fields such as [...] computer science, where the raw materials are informational and can be freely shared online." ... or we could reach the point, unthinkable only few years ago, of being able to exchange text messages in almost real time! OMG, think of the possibilities! Seriously, does the author even browse the internet?
  •  
    I do not agree with you F., you are citing out of context! Sharing messages does not make a collaboration, nor does a forum, .... You need a set of rules and a common objective. This is clearly observable in "some team", where these rules are lacking, making team work inexistent. The additional difficulties here are that it involves people that are almost strangers to each other, and the immateriality of the project. The support they are using (web, wiki) is only secondary. What they achieved is remarkable, disregarding the subject!
  •  
    I think we will just have to agree to disagree then :) Open source developers have been organizing themselves with emails since the early '90s, and most projects (e.g., the Linux kernel) still do not use anything else today. The Linux kernel mailing list gets around 400 messages per day, and they are managing just fine to scale as the number of contributors increases. I agree that what they achieved is remarkable, but it is more for "what" they achieved than "how". What they did does not remotely qualify as "massively" collaborative: again, many open source projects are managed collaboratively by thousands of people, and many of them are in the multi-million lines of code range. My personal opinion of why in the scientific world these open models are having so many difficulties is that the scientific community today is (globally, of course there are many exceptions) a closed, mostly conservative circle of people who are scared of changes. There is also the fact that the barrier of entry in a scientific community is very high, but I think that this should merely scale down the number of people involved and not change the community "qualitatively". I do not think that many research activities are so much more difficult than, e.g., writing an O(1) scheduler for an Operating System or writing a new balancing tree algorithm for efficiently storing files on a filesystem. Then there is the whole issue of scientific publishing, which, in its current form, is nothing more than a racket. No wonder traditional journals are scared to death by these open-science movements.
  •  
    here we go ... nice controversy! but maybe too many things mixed up together - open science journals vs traditional journals, conservatism of science community wrt programmers (to me one of the reasons for this might be the average age of both groups, which is probably more than 10 years apart ...) and then using emailing wrt other collaboration tools .... .... will have to look at the paper now more carefully ... (I am surprised to see no comment from José or Marek here :-)
  •  
    My point about your initial comment is that it is simplistic to infer that emails imply collaborative work. You actually use the word "organize", what does it mean indeed. In the case of Linux, what makes the project work is the rules they set and the management style (hierachy, meritocracy, review). Mailing is just a coordination mean. In collaborations and team work, it is about rules, not only about the technology you use to potentially collaborate. Otherwise, all projects would be successful, and we would noy learn management at school! They did not write they managed the colloboration exclusively because of wikipedia and emails (or other 2.0 technology)! You are missing the part that makes it successful and remarkable as a project. On his blog the guy put a list of 12 rules for this project. None are related to emails, wikipedia, forums ... because that would be lame and your comment would make sense. Following your argumentation, the tools would be sufficient for collaboration. In the ACT, we have plenty of tools, but no team work. QED
  •  
    the question on the ACT team work is one that is coming back continuously and it always so far has boiled down to the question of how much there need and should be a team project to which everybody inthe team contributes in his / her way or how much we should leave smaller, flexible teams within the team form and progress, more following a bottom-up initiative than imposing one from top-down. At this very moment, there are at least 4 to 5 teams with their own tools and mechanisms which are active and operating within the team. - but hey, if there is a real will for one larger project of the team to which all or most members want to contribute, lets go for it .... but in my view, it should be on a convince rather than oblige basis ...
  •  
    It is, though, indicative that some of the team member do not see all the collaboration and team work happening around them. We always leave the small and agile sub-teams to form and organize themselves spontaneously, but clearly this method leaves out some people (be it for their own personal attitude or be it for pure chance) For those cases which we could think to provide the possibility to participate in an alternative, more structured, team work where we actually manage the hierachy, meritocracy and perform the project review (to use Joris words).
  •  
    I am, and was, involved in "collaboration" but I can say from experience that we are mostly a sum of individuals. In the end, it is always one or two individuals doing the job, and other waiting. Sometimes even, some people don't do what they are supposed to do, so nothing happens ... this could not be defined as team work. Don't get me wrong, this is the dynamic of the team and I am OK with it ... in the end it is less work for me :) team = 3 members or more. I am personally not looking for a 15 member team work, and it is not what I meant. Anyway, this is not exactly the subject of the paper.
  •  
    My opinion about this is that a research team, like the ACT, is a group of _people_ and not only brains. What I mean is that people have feelings, hate, anger, envy, sympathy, love, etc about the others. Unfortunately(?), this could lead to situations, where, in theory, a group of brains could work together, but not the same group of people. As far as I am concerned, this happened many times during my ACT period. And this is happening now with me in Delft, where I have the chance to be in an even more international group than the ACT. I do efficient collaborations with those people who are "close" to me not only in scientific interest, but also in some private sense. And I have people around me who have interesting topics and they might need my help and knowledge, but somehow, it just does not work. Simply lack of sympathy. You know what I mean, don't you? About the article: there is nothing new, indeed. However, why it worked: only brains and not the people worked together on a very specific problem. Plus maybe they were motivated by the idea of e-collaboration. No revolution.
  •  
    Joris, maybe I made myself not clear enough, but my point was only tangentially related to the tools. Indeed, it is the original article mention of "development of new online tools" which prompted my reply about emails. Let me try to say it more clearly: my point is that what they accomplished is nothing new methodologically (i.e., online collaboration of a loosely knit group of people), it is something that has been done countless times before. Do you think that now that it is mathematicians who are doing it makes it somehow special or different? Personally, I don't. You should come over to some mailing lists of mathematical open-source software (e.g., SAGE, Pari, ...), there's plenty of online collaborative research going on there :) I also disagree that, as you say, "in the case of Linux, what makes the project work is the rules they set and the management style (hierachy, meritocracy, review)". First of all I think the main engine of any collaboration like this is the objective, i.e., wanting to get something done. Rules emerge from self-organization later on, and they may be completely different from project to project, ranging from almost anarchy to BDFL (benevolent dictator for life) style. Given this kind of variety that can be observed in open-source projects today, I am very skeptical that any kind of management rule can be said to be universal (and I am pretty sure that the overwhelming majority of project organizers never went to any "management school"). Then there is the social aspect that Tamas mentions above. From my personal experience, communities that put technical merit above everything else tend to remain very small and generally become irrelevant. The ability to work and collaborate with others is the main asset the a participant of a community can bring. I've seen many times on the Linux kernel mailing list contributions deemed "technically superior" being disregarded and not considered for inclusion in the kernel because it was clear that
  •  
    hey, just catched up the discussion. For me what is very new is mainly the framework where this collaborative (open) work is applied. I haven't seen this kind of working openly in any other field of academic research (except for the Boinc type project which are very different, because relying on non specialists for the work to be done). This raise several problems, and mainly the one of the credit, which has not really been solved as I read in the wiki (is an article is written, who writes it, what are the names on the paper). They chose to refer to the project, and not to the individual researchers, as a temporary solution... It is not so surprising for me that this type of work has been first done in the domain of mathematics. Perhaps I have an ideal view of this community but it seems that the result obtained is more important than who obtained it... In many areas of research this is not the case, and one reason is how the research is financed. To obtain money you need to have (scientific) credit, and to have credit you need to have papers with your name on it... so this model of research does not fit in my opinion with the way research is governed. Anyway we had a discussion on the Ariadnet on how to use it, and one idea was to do this kind of collaborative research; idea that was quickly abandoned...
  •  
    I don't really see much the problem with giving credit. It is not the first time a group of researchers collectively take credit for a result under a group umbrella, e.g., see Nicolas Bourbaki: http://en.wikipedia.org/wiki/Bourbaki Again, if the research process is completely transparent and publicly accessible there's no way to fake contributions or to give undue credit, and one could cite without problems a group paper in his/her CV, research grant application, etc.
  •  
    Well my point was more that it could be a problem with how the actual system works. Let say you want a grant or a position, then the jury will count the number of papers with you as a first author, and the other papers (at least in France)... and look at the impact factor of these journals. Then you would have to set up a rule for classifying the authors (endless and pointless discussions), and give an impact factor to the group...?
  •  
    it seems that i should visit you guys at estec... :-)
  •  
    urgently!! btw: we will have the ACT christmas dinner on the 9th in the evening ... are you coming?
jcunha

Silicon Valley celebrates Moore's Law 50 years anniversary - 2 views

  •  
    A bit late, but it is very interesting and instructive to listen to Gordon Moore's words "It almost doubled every year (...) so I said in the next 10 years it's going to continue to double every year, we are going to go from 16 components on a chip to 16 000. Pretty wild extrapolation!". This extrapolation (exponential with only 5 initial points) is now well-known and is one of the things that changed the World, it is pretty amazing how this "wild" futuristic vision came true.
  •  
    A great source is also the blog from Chris Mack (http://life.lithoguru.com/?p=451) who is a semicon pioneer and publisher of many books on the subject. He wrote an article for IEEE Spectrum on Moore's law and its future. Find it here, http://spectrum.ieee.org/semiconductors/processors/the-multiple-lives-of-moores-law
  •  
    Whenever I think about moore's law and extrapolating like that I end up back at this xkcd comic https://xkcd.com/605/
pacome delva

Radiation fears stalk stellar mission - 0 views

  • Concern over ESA's handling of the radiation issue caused Michael Perryman, former GAIA project scientist, to resign from the agency in 2008. But GAIA science-team member Lennart Lindegren, an astronomer at Lund Observatory in Sweden, is confident that GAIA's unprecedented accuracy will be feasible. GAIA researchers will continue to perform tests and calibrations until at least 2011, which will include irradiating the CCDs at space-like temperatures. Lindegren admits, however, that they can never be certain of success until the spacecraft is in orbit and starts sending back data.
  •  
    metamaterials to stop radiations?
  •  
    hmm ... in the optical range it is already a bit tricky .... lets brainstorm a bit about it ... anything in the literature? Luzi?
  •  
    the components of the metamaterial need to have a size inversely proportional to the wavelenght, but here we are talking about high energy protons. So the idea was: is there metamaterials that change the electrical properties to solve this problem of holes in the material... but i dont have a clue !
Joris _

SPACE.com -- Bigelow Aerospace Soars with Private Space Station Deals - 0 views

  • A private space company offering room on inflatable space habitats for research has found a robust international market
  • A question that continues to float through the halls of NASA and the Congress: Is there a commercial market for utilizing space?
pacome delva

NASA Administrator Promises Not to Cannibalize Science Budget - 0 views

  • Bolden didn't provide any direct answers to whether the agency would continue to send humans into space in the coming decade, as recommended by a report from a blue-ribbon panel last October. However, Bolden promised that the program would not be paid for through cuts to the agency's science budget, which has been cannibalized in recent years to support space flights. The auditorium erupted in applause. Bolden avoided getting into any specifics on the next generation of NASA manned spacecraft, a topic of hot debate.
Joris _

Robot mini space shuttle is go for April, says US air force * The Register - 2 views

  • concept of operations development for reusable space vehicle technologies.
  • a robotic spaceplane able to remain in orbit for substantial periods before re-entering the atmosphere and coming in for a runway landing automatically
  •  
    make you wonder when the FLPP will actually lead to something flying !
  •  
    I did not even know that the airforce continued the X37 project when it was cancelled at NASA ... interesting stuff indeed ...
nikolas smyrlakis

mentored by the Advanced Concepts Team for Google Summer of Code 2010 - 4 views

  •  
    you propably already know,I post it for the twitter account and for your comments
  • ...4 more comments...
  •  
    once again one of these initiatives that came up from a situation and that would never have been possible with a top-down approach .... fantastic! and as Dario said: we are apparently where NASA still has to go with this :-)
  •  
    Actually, NASA Ames did that already within the NASA Open Source Agreement in 2008 for a V&V software!
  •  
    indeed ... you are right .... interesting project btw - they started in 1999, were in 2005 the first NASA project on Sourceforge and won several awards .... then this entry why they did not participate last year: "05/01/09: Skipping this years Google Summer-of-Code - many of you have asked why we are not participating in this years Summer of Code. The answer is that both John and Peter were too busy with other assignments to set this up in time. We will be back in 2010. At least we were able to compensate with a limited number of NASA internships to continue some of last years projects." .... but I could not find them in this years selected list - any clue?
  •  
    but in any case, according to the apple guru, Java is a dying technology, so their project might as well ...
  •  
    They participate under the name "The Java Pathfinder Team" (http://babelfish.arc.nasa.gov/trac/jpf/wiki/events/soc2010). It is actually a very useful project for both education and industry (Airbus created a consortium on model checking soft, and there is a lot of research on it) As far as I know, TAS had some plans of using Java onboard spacecrafts, 2 years ago. Not sure the industry is really sensible about Jobs' opinions ;) particularly if there is no better alternative!
Ma Ru

MY favourite ESA project continues... - 3 views

  •  
    Anyone of you among the lucky winners?
LeopoldS

Macroscopic invisibility cloaking of visible light : Nature Communications : Nature Pub... - 3 views

  •  
    and all this without magic metamaterials ...
  • ...1 more comment...
  •  
    Wellwellwell! I don't know how I have to complain, since I could not yet read the full article, but I'm sure I will :-).
  •  
    It's funny to see how people get more and more humble in the desperate attempt to save their stupid ideas... At the beginning was the brave and bold aim to cloak something in free space (in a sphere or a cylinder). This requires inhomogeneous, anisotropic, magnetic materials; hopeless!! So one reduces to one polarization, now we have inhomogenous, anisotropic materials; still hopeless! At this point one downgraded the pretension: instead of cloaking in free space, we make a "carpet cloak" and hide an object behind an invisible dent in a mirror. But if that shall be continuous, we still need inhomogeneity and this is very hard. So now instead of a dent we take a cone and then it is claimed to work ... for ONE polarization. But of course the cloak can't work at all incident angles... irony of fate: everything is now made from birefringent media, the antithesis of what the metamaterials dogma was at the beginning!
  •  
    Hi Luzi, can you please send me the paper. We are writing a project based on sulfates and carbonates, and all this BS sounds great for the introduction (The authors used Calcite as birefringent material)
Giusi Schiavone

people control thoughts better when see thier brain activity - 1 views

  •  
    if we were able to continuously look at our brain activity we could improve our way to use the mind
Francesco Biscani

What Larry Page really needs to do to return Google to its startup roots - 0 views

  • I worked at Google from 2005-2010, and saw the company go through many changes, and a huge increase in staff.  Most importantly, I saw the company go from a place where engineers were seen as violent disruptors and innovators, to a place where doing things “The Google Way” was king, and where thinking outside the box was discouraged and even chastised.
  • Let engineers do what they do best, and forget the rest.
  • This is probably the most important single point.  Engineers at Google spend way too much time fussing about with everything other than engineering and product design.
  • ...2 more annotations...
  • Meetings.  Seriously, people are drenched in “status update” and “team” meetings.
  • Weekly Snippets, perf, etc. I was continually amazed by the amount of “extra cruft work” that goes on.  I know it sounds important, but engineers should be coding & designing.
  •  
    An opinion piece by an ex-Googler, talking, among other things, about how corporate culture is creeping into Google. I've highlighted some snippets which I've found eerily familiar :)
Luís F. Simões

Following SpaceX down the rabbit hole -- The Space Review - 4 views

  • He then went on to remind the press that his company’s goal is to continue to lower the cost of access to space because high launch costs were “the fundamental factor preventing humanity from becoming a spacefaring civilization.”
  • First, two Falcon Heavy launches could field a return to the Moon
  • Second, a single Falcon Heavy could launch a Mars sample return mission.
  • ...3 more annotations...
  • Also possible with a single launch: a first ever human mission to an asteroid.
  • what effect Falcon Heavy might have on the costs of supporting the International Space Station. The ability to launch twice the supply capacity provided by the shuttle at something on the order of 20 percent of the cost changes the calculus entirely. So much so in fact, it opens the door for contemplating an entirely different future for ISS in which it never follows Mir into the Pacific.
  • Following conclusions offered by the Augustine Committee, the Obama Administration cancelled Project Constellation as unaffordable under existing budget limits, and supported instead the ambling, but cost-contained “flexible path.” If the Falcon Heavy is available, however, the rationale for selecting the flexible path—because it’s the only thing we can afford—simply doesn’t exist.
  •  
    an analysis of implications of the Falcon Heavy announcement
Lionel Jacques

The end of GMT ? - 3 views

  •  
    Greenwich could lose its place at the centre of global time if a move to "atomic time" is voted in by the International Telecommunication Union in Geneva in January 2012.
  • ...1 more comment...
  •  
    The article says it can lead to abandoning the Daylight Wasting Time in winter, so if that's the case, I'm definitely for.
  •  
    Haha this is really a British article... Already since 1972 we don't use GMT but UTC, which is based on atomic clocks. However British continue to call it GMT... The question is to drop the leap second in UTC, and France is definitely for this change (for scientific motives of course...;) I don't see how this is connected to winter time however... And they shouldn't worry Greenwich is still the beginning of the world with 0 degree longitude !
  •  
    "the end of GMT as an international standard could accelerate the move to keep British Summer Time into the winter, letting us have lighter evenings." As I understand it, if GMT looses its "prestigious" status, then it would be easier to push through all-year BST in UK.
Luís F. Simões

Lust in space: Russians lose control of gecko sex satellite | Al Jazeera America - 5 views

  • Lizards were sent into orbit as part of study into effects of weightlessness on sexual intercourse
  • On Thursday, the team behind the research confirmed that the vessel was not responding to commands, potentially leaving the reptiles to their out-of-this-world sexual intercourse while video footage continues to beam down to Earth.
  •  
    I still think, the lizards have evolved at an unexpectedly high rate and have now taken over the satellite...
‹ Previous 21 - 40 of 43 Next ›
Showing 20 items per page