Skip to main content

Home/ Advanced Concepts Team/ Group items tagged discussion

Rss Feed Group items tagged

tvinko

Massively collaborative mathematics : Article : Nature - 28 views

  •  
    peer-to-peer theorem-proving
  • ...14 more comments...
  •  
    Or: mathematicians catch up with open-source software developers :)
  •  
    "Similar open-source techniques could be applied in fields such as [...] computer science, where the raw materials are informational and can be freely shared online." ... or we could reach the point, unthinkable only few years ago, of being able to exchange text messages in almost real time! OMG, think of the possibilities! Seriously, does the author even browse the internet?
  •  
    I do not agree with you F., you are citing out of context! Sharing messages does not make a collaboration, nor does a forum, .... You need a set of rules and a common objective. This is clearly observable in "some team", where these rules are lacking, making team work inexistent. The additional difficulties here are that it involves people that are almost strangers to each other, and the immateriality of the project. The support they are using (web, wiki) is only secondary. What they achieved is remarkable, disregarding the subject!
  •  
    I think we will just have to agree to disagree then :) Open source developers have been organizing themselves with emails since the early '90s, and most projects (e.g., the Linux kernel) still do not use anything else today. The Linux kernel mailing list gets around 400 messages per day, and they are managing just fine to scale as the number of contributors increases. I agree that what they achieved is remarkable, but it is more for "what" they achieved than "how". What they did does not remotely qualify as "massively" collaborative: again, many open source projects are managed collaboratively by thousands of people, and many of them are in the multi-million lines of code range. My personal opinion of why in the scientific world these open models are having so many difficulties is that the scientific community today is (globally, of course there are many exceptions) a closed, mostly conservative circle of people who are scared of changes. There is also the fact that the barrier of entry in a scientific community is very high, but I think that this should merely scale down the number of people involved and not change the community "qualitatively". I do not think that many research activities are so much more difficult than, e.g., writing an O(1) scheduler for an Operating System or writing a new balancing tree algorithm for efficiently storing files on a filesystem. Then there is the whole issue of scientific publishing, which, in its current form, is nothing more than a racket. No wonder traditional journals are scared to death by these open-science movements.
  •  
    here we go ... nice controversy! but maybe too many things mixed up together - open science journals vs traditional journals, conservatism of science community wrt programmers (to me one of the reasons for this might be the average age of both groups, which is probably more than 10 years apart ...) and then using emailing wrt other collaboration tools .... .... will have to look at the paper now more carefully ... (I am surprised to see no comment from José or Marek here :-)
  •  
    My point about your initial comment is that it is simplistic to infer that emails imply collaborative work. You actually use the word "organize", what does it mean indeed. In the case of Linux, what makes the project work is the rules they set and the management style (hierachy, meritocracy, review). Mailing is just a coordination mean. In collaborations and team work, it is about rules, not only about the technology you use to potentially collaborate. Otherwise, all projects would be successful, and we would noy learn management at school! They did not write they managed the colloboration exclusively because of wikipedia and emails (or other 2.0 technology)! You are missing the part that makes it successful and remarkable as a project. On his blog the guy put a list of 12 rules for this project. None are related to emails, wikipedia, forums ... because that would be lame and your comment would make sense. Following your argumentation, the tools would be sufficient for collaboration. In the ACT, we have plenty of tools, but no team work. QED
  •  
    the question on the ACT team work is one that is coming back continuously and it always so far has boiled down to the question of how much there need and should be a team project to which everybody inthe team contributes in his / her way or how much we should leave smaller, flexible teams within the team form and progress, more following a bottom-up initiative than imposing one from top-down. At this very moment, there are at least 4 to 5 teams with their own tools and mechanisms which are active and operating within the team. - but hey, if there is a real will for one larger project of the team to which all or most members want to contribute, lets go for it .... but in my view, it should be on a convince rather than oblige basis ...
  •  
    It is, though, indicative that some of the team member do not see all the collaboration and team work happening around them. We always leave the small and agile sub-teams to form and organize themselves spontaneously, but clearly this method leaves out some people (be it for their own personal attitude or be it for pure chance) For those cases which we could think to provide the possibility to participate in an alternative, more structured, team work where we actually manage the hierachy, meritocracy and perform the project review (to use Joris words).
  •  
    I am, and was, involved in "collaboration" but I can say from experience that we are mostly a sum of individuals. In the end, it is always one or two individuals doing the job, and other waiting. Sometimes even, some people don't do what they are supposed to do, so nothing happens ... this could not be defined as team work. Don't get me wrong, this is the dynamic of the team and I am OK with it ... in the end it is less work for me :) team = 3 members or more. I am personally not looking for a 15 member team work, and it is not what I meant. Anyway, this is not exactly the subject of the paper.
  •  
    My opinion about this is that a research team, like the ACT, is a group of _people_ and not only brains. What I mean is that people have feelings, hate, anger, envy, sympathy, love, etc about the others. Unfortunately(?), this could lead to situations, where, in theory, a group of brains could work together, but not the same group of people. As far as I am concerned, this happened many times during my ACT period. And this is happening now with me in Delft, where I have the chance to be in an even more international group than the ACT. I do efficient collaborations with those people who are "close" to me not only in scientific interest, but also in some private sense. And I have people around me who have interesting topics and they might need my help and knowledge, but somehow, it just does not work. Simply lack of sympathy. You know what I mean, don't you? About the article: there is nothing new, indeed. However, why it worked: only brains and not the people worked together on a very specific problem. Plus maybe they were motivated by the idea of e-collaboration. No revolution.
  •  
    Joris, maybe I made myself not clear enough, but my point was only tangentially related to the tools. Indeed, it is the original article mention of "development of new online tools" which prompted my reply about emails. Let me try to say it more clearly: my point is that what they accomplished is nothing new methodologically (i.e., online collaboration of a loosely knit group of people), it is something that has been done countless times before. Do you think that now that it is mathematicians who are doing it makes it somehow special or different? Personally, I don't. You should come over to some mailing lists of mathematical open-source software (e.g., SAGE, Pari, ...), there's plenty of online collaborative research going on there :) I also disagree that, as you say, "in the case of Linux, what makes the project work is the rules they set and the management style (hierachy, meritocracy, review)". First of all I think the main engine of any collaboration like this is the objective, i.e., wanting to get something done. Rules emerge from self-organization later on, and they may be completely different from project to project, ranging from almost anarchy to BDFL (benevolent dictator for life) style. Given this kind of variety that can be observed in open-source projects today, I am very skeptical that any kind of management rule can be said to be universal (and I am pretty sure that the overwhelming majority of project organizers never went to any "management school"). Then there is the social aspect that Tamas mentions above. From my personal experience, communities that put technical merit above everything else tend to remain very small and generally become irrelevant. The ability to work and collaborate with others is the main asset the a participant of a community can bring. I've seen many times on the Linux kernel mailing list contributions deemed "technically superior" being disregarded and not considered for inclusion in the kernel because it was clear that
  •  
    hey, just catched up the discussion. For me what is very new is mainly the framework where this collaborative (open) work is applied. I haven't seen this kind of working openly in any other field of academic research (except for the Boinc type project which are very different, because relying on non specialists for the work to be done). This raise several problems, and mainly the one of the credit, which has not really been solved as I read in the wiki (is an article is written, who writes it, what are the names on the paper). They chose to refer to the project, and not to the individual researchers, as a temporary solution... It is not so surprising for me that this type of work has been first done in the domain of mathematics. Perhaps I have an ideal view of this community but it seems that the result obtained is more important than who obtained it... In many areas of research this is not the case, and one reason is how the research is financed. To obtain money you need to have (scientific) credit, and to have credit you need to have papers with your name on it... so this model of research does not fit in my opinion with the way research is governed. Anyway we had a discussion on the Ariadnet on how to use it, and one idea was to do this kind of collaborative research; idea that was quickly abandoned...
  •  
    I don't really see much the problem with giving credit. It is not the first time a group of researchers collectively take credit for a result under a group umbrella, e.g., see Nicolas Bourbaki: http://en.wikipedia.org/wiki/Bourbaki Again, if the research process is completely transparent and publicly accessible there's no way to fake contributions or to give undue credit, and one could cite without problems a group paper in his/her CV, research grant application, etc.
  •  
    Well my point was more that it could be a problem with how the actual system works. Let say you want a grant or a position, then the jury will count the number of papers with you as a first author, and the other papers (at least in France)... and look at the impact factor of these journals. Then you would have to set up a rule for classifying the authors (endless and pointless discussions), and give an impact factor to the group...?
  •  
    it seems that i should visit you guys at estec... :-)
  •  
    urgently!! btw: we will have the ACT christmas dinner on the 9th in the evening ... are you coming?
Luís F. Simões

How to Grow a Mind: Statistics, Structure, and Abstraction - 4 views

  •  
    a nice review on the wonders of Hierarchical Bayesian models. It cites a paper on probabilistic programming languages that might be relevant given our recent discussions. At Hippo's farewell lunch there was a discussion on how kids are able to learn something as complex as language from a limited amount of observations, while Machine Learning algorithms no matter how many millions of instances you throw at them, don't learn beyond some point. If that subject interested you, you might like this paper.
  •  
    Had an opportunity to listen to JBT and TLG during one summer school.. if they're half as good in writing as they are in speaking, should be a decent read...
ESA ACT

ACPD - Papers in Open Discussion - 0 views

  •  
    Useful journal for the upcoming Earth Systems Science. For geoengineering, check out this: http://www.atmos-chem-phys-discuss.net/9/2559/2009/acpd-9-2559-2009.pdf (currently first on the list of this bookmark)
LeopoldS

Global Innovation Commons - 4 views

  •  
    nice initiative!
  • ...6 more comments...
  •  
    Any viral licence is a bad license...
  •  
    I'm pretty confident I'm about to open a can of worms, but mind explaining why? :)
  •  
    I am less worried about the can of worms ... actually eager to open it ... so why????
  •  
    Well, the topic GPL vs other open-source licenses (e.g., BSD, MIT, etc.) is old as the internet and it has provided material for long and glorious flame wars. The executive summary is that the GPL license (the one used by Linux) is a license which imposes some restrictions on the way you are allowed to (re)use the code. Specifically, if you re-use or modify GPL code and re-distribute it, you are required to make it available again under the GPL license. It is called "viral" because once you use a bit of GPL code, you are required to make the whole application GPL - so in this sense GPL code replicates like a virus. On the other side of the spectrum, there are the so-called BSD-like licenses which have more relaxed requirements. Usually, the only obligation they impose is to acknowledge somewhere (e.g., in a README file) that you have used some BSD code and who wrote it (this is called "attribution clause"), but they do not require to re-distribute the whole application under the same license. GPL critics usually claim that the license is not really "free" because it does not allow you to do whatever you want with the code without restrictions. GPL proponents claim that the requirements imposed by the GPL are necessary to safeguard the freedom of the code, in order to avoid being able to re-use GPL code without giving anything back to the community (which the BSD license allow: early versions of Microsoft Windows, for instance, had the networking code basically copy-pasted from BSD-licensed versions of Unix). In my opinion (and this point is often brought up in the debates) the division pro/against GPL mirrors somehow the division between anti/pro anarchism. Anarchists claim that the only way to be really free is the absence of laws, while non-anarchist maintain that the only practical way to be free is to have laws (which by definition limit certain freedoms). So you can see how the topic can quickly become inflammatory :) GPL at the current time is used by aro
  •  
    whoa, the comment got cut off. Anyway, I was just saying that at the present time the GPL license is used by around 65% of open source projects, including the Linux kernel, KDE, Samba, GCC, all the GNU utils, etc. The topic is much deeper than this brief summary, so if you are interested in it, Leopold, we can discuss it at length in another place.
  •  
    Thanks for the record long comment - am sure that this is longest ever made to an ACT diigo post! On the topic, I would rather lean for the GPL license (which I also advocated for the Marek viewer programme we put on source forge btw), mainly because I don't trust that open source is by nature delivering a better product and thus will prevail but I still would like to succeed, which I am not sure it would if there were mainly BSD like licenses around. ... but clearly, this is an outsider talking :-)
  •  
    btw: did not know the anarchist penchant of Marek :-)
  •  
    Well, not going into the discussion about GPL/BSD, the viral license in this particular case in my view simply undermines the "clean and clear" motivations of the initiative authors - why should *they* be credited for using something they have no rights for? And I don't like viral licences because they prevent using things released under this licence to all those people who want to release their stuff under a different licence, thus limiting the usefulness of the stuff released on that licence :) BSD is not a perfect license too, it also had major flaws And I'm not an anarchist, lol
johannessimon81

Nuclear isomer - 2 views

  •  
    We had a short discussion yesterday about using nuclear isomers as batteries for spacecraft. The principle is that energy is stored as an excitation of the nucleus which can then release the energy as a gamma-photon. However angular momentum has to be conserved an this suppresses the decay strongly - making these states stable up to 10^35 longer than a typical decay.
  •  
    The key is the triggering of the dacay wheras the triggering comsumes less energy than the decay provides. The x-ray based triggering of the gamma photon decay turned out to be quite controversial and needs significantly more scientific attention.
  •  
    I think that there should be some references and traces of the discussions we had on this on the shared drive or wiki ... One other aspect: converting the omnidirectional gamma bursts into useful energy ....
LeopoldS

Yahoo Orders Home Workers Back to the Office - NYTimes.com - 0 views

  •  
    interesting discussion on benefits and drawbacks of teleworking ...
LeopoldS

Software » OpenStack Open Source Cloud Computing Software - 2 views

  •  
    for the ACT? already posted here once I think but can#t remember the discussion
LeopoldS

On the Habitable Zones of Circumbinary Planetary Systems - Abstract - The Astrophysical... - 2 views

  •  
    remember our recent discussion about this?
LeopoldS

Schumpeter: More than just a game | The Economist - 3 views

  •  
    remember the discussion I tried to trigger in the team a few weeks ago ...
  • ...5 more comments...
  •  
    main quote I take from the article: "gamification is really a cover for cynically exploiting human psychology for profit"
  •  
    I would say that it applies to management in general :-)
  •  
    which is exactly why it will never work .... and surprisingly "managers" fail to understand this very simple fact.
  •  
    ... "gamification is really a cover for cynically exploiting human psychology for profit" --> "Why Are Half a Million People Poking This Giant Cube?" http://www.wired.com/gamelife/2012/11/curiosity/
  •  
    I think the "essence" of the game is its uselessness... workers need exactly the inverse, to find a meaning in what they do !
  •  
    I love the linked article provided by Johannes! It expresses very elegantly why I still fail to understand even extremely smart and busy people in my view apparently waiting their time in playing computer games - but I recognise that there is something in games that we apparently need / gives us something we cherish .... "In fact, half a million players so far have registered to help destroy the 64 billion tiny blocks that compose that one gigantic cube, all working in tandem toward a singular goal: discovering the secret that Curiosity's creator says awaits one lucky player inside. That's right: After millions of man-hours of work, only one player will ever see the center of the cube. Curiosity is the first release from 22Cans, an independent game studio founded earlier this year by Peter Molyneux, a longtime game designer known for ambitious projects like Populous, Black & White and Fable. Players can carve important messages (or shameless self-promotion) onto the face of the cube as they whittle it to nothing. Image: Wired Molyneux is equally famous for his tendency to overpromise and under-deliver on his games. In 2008, he said that his upcoming game would be "such a significant scientific achievement that it will be on the cover of Wired." That game turned out to be Milo & Kate, a Kinect tech demo that went nowhere and was canceled. Following this, Molyneux left Microsoft to go indie and form 22Cans. Not held back by the past, the Molyneux hype train is going full speed ahead with Curiosity, which the studio grandiosely promises will be merely the first of 22 similar "experiments." Somehow, it is wildly popular. The biggest challenge facing players of Curiosity isn't how to blast through the 2,000 layers of the cube, but rather successfully connecting to 22Cans' servers. So many players are attempting to log in that the server cannot handle it. Some players go for utter efficiency, tapping rapidly to rack up combo multipliers and get more
  •  
    why are video games so much different than collecting stamps or spotting birds or planes ? One could say they are all just hobbies
Dario Izzo

Stacked Approximated Regression Machine: A Simple Deep Learning Approach - 5 views

  •  
    from one of the reddit threads discussing this: "bit fishy, crazy if real". "Incredible claims: - Train only using about 10% of imagenet-12, i.e. around 120k images (i.e. they use 6k images per arm) - get to the same or better accuracy as the equivalent VGG net - Training is not via backprop but more simpler PCA + Sparsity regime (see section 4.1), shouldn't take more than 10 hours just on CPU probably "
  •  
    clicking the link says the manuscript was withdrawn :))
  •  
    This "one-shot learning" paper by Googe Deepmind also claims to be able to learn from very few training data. Thought it might be interesting for you guys: https://arxiv.org/pdf/1605.06065v1.pdf
Luís F. Simões

Our approach to replication in computational science - 2 views

  • So what did we do to make this paper extra super replicable? If you go to the paper Web site, you'll find:
  • p.s. I think I have to refer to this cancer results not reproducible paper somewhere. Done.
  •  
    good discussion on the replicability/reproducibility of scientific results (also a nice example of how to do it right... in bioinformatics at least)
LeopoldS

NPS CUBESAT LAUNCHER DESIGN, PROCESS AND REQUIREMENTS - 2 views

  •  
    Interesting mater thesis that lead to the just announced new launcher ! Remember our discussion after the advanced concepts workshop!?!?!
  •  
    do you know about this? https://www.qb50.eu/launch.php
  •  
    I knew about the project only, but not about this very interesting missile ... "The launcher has a remarkable track record of several hundred successful launches and only 1 failure. Shtil is marketed by the State Rocket Center Makeyev."
jmlloren

Executable Paper Grand Challenge - 4 views

  •  
    It is not new, but I can't remember any discussion on this concept at the ACT
Luís F. Simões

HP Dreams of Internet Powered by Phone Chips (And Cow Chips) | Wired.com - 0 views

  • For Hewlett Packard Fellow Chandrakat Patel, there’s a “symbiotic relationship between IT and manure.”
  • Patel is an original thinker. He’s part of a group at HP Labs that has made energy an obsession. Four months ago, Patel buttonholed former Federal Reserve Chairman Alan Greenspan at the Aspen Ideas Festival to sell him on the idea that the joule should be the world’s global currency.
  • Data centers produce a lot of heat, but to energy connoisseurs it’s not really high quality heat. It can’t boil water or power a turbine. But one thing it can do is warm up poop. And that’s how you produce methane gas. And that’s what powers Patel’s data center. See? A symbiotic relationship.
  • ...1 more annotation...
  • Financial house Cantor Fitzgerald is interested in Project Moonshot because it thinks HP’s servers may have just what it takes to help the company’s traders understand long-term market trends. Director of High-Frequency Trading Niall Dalton says that while the company’s flagship trading platform still needs the quick number-crunching power that comes with the powerhog chips, these low-power Project Moonshot systems could be great for analyzing lots and lots of data — taking market data from the past three years, for example, and running a simulation.
  •  
    of relevance to this discussion: Koomey's Law, a Moore's Law equivalent for computing's energetic efficiency http://www.economist.com/node/21531350 http://hardware.slashdot.org/story/11/09/13/2148202/whither-moores-law-introducing-koomeys-law
Isabelle Dicaire

In the zone: How scientists search for habitable planets - 1 views

  •  
    General discussion on how to search for habitable planets
Tom Gheysens

Super-material shrugs off molten metal : Nature News & Comment - 0 views

  •  
    these are the latest in repellent surfaces.  a very nice video and as we previously discussed could be very useful in the ISS
Tom Gheysens

Computer searches web 24/7 to analyze images and teach itself common sense - 0 views

  •  
    Now this is a step in the right direction of the discussion we had in one of the wednesday meetings "thoughts of a biologist part 1" :)
Alexander Wittig

Picture This: NVIDIA GPUs Sort Through Tens of Millions of Flickr Photos - 2 views

  •  
    Strange and exotic cityscapes. Desolate wilderness areas. Dogs that look like wookies. Flickr, one of the world's largest photo sharing services, sees it all. And, now, Flickr's image recognition technology can categorize more than 11 billion photos like these. And it does it automatically. It's called "Magic View." Magical deep learning! Buzzword attack!
  • ...4 more comments...
  •  
    and here comes my standard question: how can we use this for space? fast detection of natural disasters onboard?
  •  
    Even on ground. You could for example teach it what nuclear reactors or missiles or other weapons you don't want look like on satellite pictures and automatically scan the world for them (basically replacing intelligence analysts).
  •  
    In fact, I think this could make a nice ACT project: counting seals from satellite imagery is an actual (and quite recent) thing: http://journals.plos.org/plosone/article?id=10.1371/journal.pone.0092613 In this publication they did it manually from a GeoEye 1 b/w image, which sounds quite tedious. Maybe one can train one of those image recognition algorithms to do it automatically. Or maybe it's a bit easier to count larger things, like elephants (also a thing).
  •  
    In HiPEAC (High Performance, embedded architecture and computation) conference I attended in the beginning of this year there was a big trend of CUDA GPU vs FPGA for hardware accelerated image processing. Most of it orbitting around discussing who was faster and cheaper with people from NVIDIA in one side and people from Xilinx and Intel in the other. I remember of talking with an IBM scientist working on hardware accelerated data processing working together with the Radio telescope institute in Netherlands about the solution where they working on (GPU CUDA). I gathered that NVIDIA GPU suits best in applications that somehow do not rely in hardware, having the advantage of being programmed in a 'easy' way accessible to a scientist. FPGA's are highly reliable components with the advantage of being available in radhard versions, but requiring specific knowledge of physical circuit design and tailored 'harsh' programming languages. I don't know what is the level of rad hardness in NVIDIA's GPUs... Therefore FPGAs are indeed the standard choice for image processing in space missions (a talk with the microelectronics department guys could expand on this), whereas GPUs are currently used in some ground based (radio astronomy or other types of telescopes). I think that on for a specific purpose as the one you mentioned, this FPGA vs GPU should be assessed first before going further.
  •  
    You're forgetting power usage. GPUs need 1000 hamster wheels worth of power while FPGAs can run on a potato. Since space applications are highly power limited, putting any kind of GPU monster in orbit or on a rover is failed idea from the start. Also in FPGAs if a gate burns out from radiation you can just reprogram around it. Looking for seals offline in high res images is indeed definitely a GPU task.... for now.
  •  
    The discussion of how to make FPGA hardware acceleration solutions easier to use for the 'layman' is starting btw http://reconfigurablecomputing4themasses.net/.
LeopoldS

SpaceX Undercut Competition to Clinch Head turning Iridium Deal | SpaceNews.com - 0 views

  •  
    very nice success by what we called in 2003 "the second space age" ... for Joris: (related to the discussion we had the other day) But the spacecraft will be assembled, integrated and tested in the United States and will include U.S. hardware, meaning they would not be allowed for launch aboard Indian or Chinese rockets under current U.S. regulations.
Nina Nadine Ridder

Arguments from Global Warming Skeptics and what the science really says - 1 views

  •  
    I think I posted this before (I think it was abt the iPod app) but regarding our discussion this afternoon it s a nice summary of the most frequent sceptic arguments...
  •  
    It's not easy not to be a global warming sceptic in a places like Noordwijk or Plymouth...
1 - 20 of 114 Next › Last »
Showing 20 items per page