<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
		>
<channel>
	<title>Comments on: Fun and Games</title>
	<atom:link href="http://www.linux-mag.com/id/7235/feed/" rel="self" type="application/rss+xml" />
	<link>http://www.linux-mag.com/id/7235/</link>
	<description>Open Source, Open Standards</description>
	<lastBuildDate>Sat, 05 Oct 2013 13:48:18 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.1</generator>
	<item>
		<title>By: a.thyssen</title>
		<link>http://www.linux-mag.com/id/7235/#comment-6139</link>
		<dc:creator>a.thyssen</dc:creator>
		<pubDate>Wed, 30 Nov -0001 00:00:00 +0000</pubDate>
		<guid isPermaLink="false">http://www.linux-mag.com/id/7235/#comment-6139</guid>
		<description>This article reminds me of another I read many years ago called,&lt;br /&gt;
&lt;a href=&quot;http://www.ubiq.com/hypertext/weiser/acmfuture2endnote.htm&quot; title=&quot;The Coming Age of Calm Technology&quot; rel=&quot;nofollow&quot;&gt;  and &lt;a href=&quot;http://www.ubiq.com/hypertext/weiser/UbiHome.html&quot; title=&quot;Ubiquitous Computing&quot; rel=&quot;nofollow&quot;&gt;.   They thought of it as the oppisite.  rather than putting people in a virtual reality to control or monitor the computer, the computer was made to control something in the real world, that people could control and study.  Works out to roughly the same goal.&lt;br /&gt;
&lt;br /&gt;
One of the results of which was a &#039;&lt;a HREF=&quot;http://www.isi.edu/~johnh/SOFTWARE/LAVAPS/index.html&quot; title=&quot;lava lamp&quot; rel=&quot;nofollow&quot;&gt;&#039; that sits on the desk top.  The lamp showed multi-colored blobs that moved and distorted in the window bottle.  The color was a hash of the process name so was always the same color for some particular process. the size of the blob was its memory foot-print relative to the other blobs, and the speed it moved represented how much CPU that process was using.&lt;br /&gt;
&lt;br /&gt;
On a quiet machine you have a very nice &#039;calm&#039; lamp that looks pretty, with slow blobs of different sizes. But when something is happening, you get large or fast moving blobs that get your attention.   Starting up a Web Browser for example creates a big blob,  while running a intensive data mining process created a BIG fast moving blob that seemed to shrink all the other processes to tiny specks.  Either way you knew something was happening, even if the computer remained responsive!&lt;br /&gt;
&lt;br /&gt;
basically it remains unobtrusive until needed. A quick glance shows that all is well! Better that having to live in a &#039;doom&#039; world all the time, just to keep the monsters from fighting each other!</description>
		<content:encoded><![CDATA[<p>This article reminds me of another I read many years ago called,<br />
<a href="http://www.ubiq.com/hypertext/weiser/acmfuture2endnote.htm" title="The Coming Age of Calm Technology" rel="nofollow">  and </a><a href="http://www.ubiq.com/hypertext/weiser/UbiHome.html" title="Ubiquitous Computing" rel="nofollow">.   They thought of it as the oppisite.  rather than putting people in a virtual reality to control or monitor the computer, the computer was made to control something in the real world, that people could control and study.  Works out to roughly the same goal.</p>
<p>One of the results of which was a &#8216;</a><a HREF="http://www.isi.edu/~johnh/SOFTWARE/LAVAPS/index.html" title="lava lamp" rel="nofollow">&#8216; that sits on the desk top.  The lamp showed multi-colored blobs that moved and distorted in the window bottle.  The color was a hash of the process name so was always the same color for some particular process. the size of the blob was its memory foot-print relative to the other blobs, and the speed it moved represented how much CPU that process was using.</p>
<p>On a quiet machine you have a very nice &#8216;calm&#8217; lamp that looks pretty, with slow blobs of different sizes. But when something is happening, you get large or fast moving blobs that get your attention.   Starting up a Web Browser for example creates a big blob,  while running a intensive data mining process created a BIG fast moving blob that seemed to shrink all the other processes to tiny specks.  Either way you knew something was happening, even if the computer remained responsive!</p>
<p>basically it remains unobtrusive until needed. A quick glance shows that all is well! Better that having to live in a &#8216;doom&#8217; world all the time, just to keep the monsters from fighting each other!</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: a.thyssen</title>
		<link>http://www.linux-mag.com/id/7235/#comment-6140</link>
		<dc:creator>a.thyssen</dc:creator>
		<pubDate>Wed, 30 Nov -0001 00:00:00 +0000</pubDate>
		<guid isPermaLink="false">http://www.linux-mag.com/id/7235/#comment-6140</guid>
		<description>My h refs in the above seemed to have failed leaving blank spots.&lt;br /&gt;
these were&lt;br /&gt;
&lt;br /&gt;
THE COMING AGE OF CALM TECHNOLOGY&lt;br /&gt;
http://www.ubiq.com/hypertext/weiser/acmfuture2endnote.htm&lt;br /&gt;
&lt;br /&gt;
Ubiquitous Computing&lt;br /&gt;
http://www.ubiq.com/hypertext/weiser/UbiHome.html&lt;br /&gt;
&lt;br /&gt;
and LavaPS  a Lava Lamp Monitor&lt;br /&gt;
http://www.isi.edu/~johnh/SOFTWARE/LAVAPS/index.html</description>
		<content:encoded><![CDATA[<p>My h refs in the above seemed to have failed leaving blank spots.<br />
these were</p>
<p>THE COMING AGE OF CALM TECHNOLOGY<br />
<a href="http://www.ubiq.com/hypertext/weiser/acmfuture2endnote.htm" rel="nofollow">http://www.ubiq.com/hypertext/weiser/acmfuture2endnote.htm</a></p>
<p>Ubiquitous Computing<br />
<a href="http://www.ubiq.com/hypertext/weiser/UbiHome.html" rel="nofollow">http://www.ubiq.com/hypertext/weiser/UbiHome.html</a></p>
<p>and LavaPS  a Lava Lamp Monitor<br />
<a href="http://www.isi.edu/~johnh/SOFTWARE/LAVAPS/index.html" rel="nofollow">http://www.isi.edu/~johnh/SOFTWARE/LAVAPS/index.html</a></p>
]]></content:encoded>
	</item>
	<item>
		<title>By: laytonjb</title>
		<link>http://www.linux-mag.com/id/7235/#comment-6141</link>
		<dc:creator>laytonjb</dc:creator>
		<pubDate>Wed, 30 Nov -0001 00:00:00 +0000</pubDate>
		<guid isPermaLink="false">http://www.linux-mag.com/id/7235/#comment-6141</guid>
		<description>Doug makes a point in this article that I think is subtle but very important. The idea is that we need to make monitoring much, much better. In fact, I will go as far as to say that monitoring runs neck and neck with programming as the biggest problem facing HPC (clusters in particular).&lt;br /&gt;
&lt;br /&gt;
Monitoring in this case is not just monitoring the nodes from a sysadmin perspective, but also monitoring user&#039;s behavior and tendencies as well as the progress of computations, etc. However, having been a sysadmin, I have a tendency to look at that aspect first. &lt;br /&gt;
&lt;br /&gt;
One of the first things we as a community could do, is to develop a set of monitoring tools that are really more &quot;push&quot; than &quot;pull&quot;. The idea is that the compute nodes monitor themselves and when there is a substantive change in whatever their are monitoring, they push only that value to the master node or to a monitoring node. This cuts down on traffic, etc. It might increase OS jitter a bit since the nodes have to be self-monitoring to some extent, which probably means daemons waking up and doing something peridocially. But I think overall it makes life a bit easier. &lt;br /&gt;
&lt;br /&gt;
For example, why not put a simple USB stick in the front USB port of each node. Then the monitoring data is written to the stick fairly quickly. You could even stash /var/log/messages on the stick. If the node fails or if a job hangs, you could either pull the data from the stick over the network, or just go grab the stick, swap in a fresh stick, and then read the data on the stick for any clues (but all sysadmin&#039;s know that if a node fails, it usually doesn&#039;t leave a trail of breadcrumbs that are easy to follow). You could even dedicated a single core to do nothing but monitor the node (this might be more palatable if we increase the core count yet again - but don&#039;t blink that&#039;s coming later this year from both AMD and Intel).&lt;br /&gt;
&lt;br /&gt;
There are other aspects to monitoring as well. Doug&#039;s comments about visualizing the results of a computation while it&#039;s still running has been done in the aerospace world for several years. There is a package called PV3 that allows you to send back solutions from a running job and examine them. This way you can tell if a job seems to be running well or not. You can also have the running job dump the current solution periodically and then you can post-process it to examine the current solution (yep - this requires a pretty decent IO subsystem as well as a good post-processing computational capabilities and network). In the aerospace world they&#039;ve adopted the phrase, &quot;computation steering&quot;. &lt;br /&gt;
&lt;br /&gt;
Lots of cool things to be done in this area. It&#039;s just not sexy enough to get funding from a gov. agency and most companies won&#039;t fund it either. It&#039;s not going to improve the speed of your applications to any noticeable degree. But it can improve the productivity and reduce some of the administration challenges associated with clusters. I most definite worthwhile effort in my opinion but unlikely to be done by anyone.&lt;br /&gt;
&lt;br /&gt;
Regardless - thanks Doug for the reminder that we need to think more creatively about how to approach clusters.&lt;br /&gt;
&lt;br /&gt;
Jeff</description>
		<content:encoded><![CDATA[<p>Doug makes a point in this article that I think is subtle but very important. The idea is that we need to make monitoring much, much better. In fact, I will go as far as to say that monitoring runs neck and neck with programming as the biggest problem facing HPC (clusters in particular).</p>
<p>Monitoring in this case is not just monitoring the nodes from a sysadmin perspective, but also monitoring user&#8217;s behavior and tendencies as well as the progress of computations, etc. However, having been a sysadmin, I have a tendency to look at that aspect first. </p>
<p>One of the first things we as a community could do, is to develop a set of monitoring tools that are really more &#8220;push&#8221; than &#8220;pull&#8221;. The idea is that the compute nodes monitor themselves and when there is a substantive change in whatever their are monitoring, they push only that value to the master node or to a monitoring node. This cuts down on traffic, etc. It might increase OS jitter a bit since the nodes have to be self-monitoring to some extent, which probably means daemons waking up and doing something peridocially. But I think overall it makes life a bit easier. </p>
<p>For example, why not put a simple USB stick in the front USB port of each node. Then the monitoring data is written to the stick fairly quickly. You could even stash /var/log/messages on the stick. If the node fails or if a job hangs, you could either pull the data from the stick over the network, or just go grab the stick, swap in a fresh stick, and then read the data on the stick for any clues (but all sysadmin&#8217;s know that if a node fails, it usually doesn&#8217;t leave a trail of breadcrumbs that are easy to follow). You could even dedicated a single core to do nothing but monitor the node (this might be more palatable if we increase the core count yet again &#8211; but don&#8217;t blink that&#8217;s coming later this year from both AMD and Intel).</p>
<p>There are other aspects to monitoring as well. Doug&#8217;s comments about visualizing the results of a computation while it&#8217;s still running has been done in the aerospace world for several years. There is a package called PV3 that allows you to send back solutions from a running job and examine them. This way you can tell if a job seems to be running well or not. You can also have the running job dump the current solution periodically and then you can post-process it to examine the current solution (yep &#8211; this requires a pretty decent IO subsystem as well as a good post-processing computational capabilities and network). In the aerospace world they&#8217;ve adopted the phrase, &#8220;computation steering&#8221;. </p>
<p>Lots of cool things to be done in this area. It&#8217;s just not sexy enough to get funding from a gov. agency and most companies won&#8217;t fund it either. It&#8217;s not going to improve the speed of your applications to any noticeable degree. But it can improve the productivity and reduce some of the administration challenges associated with clusters. I most definite worthwhile effort in my opinion but unlikely to be done by anyone.</p>
<p>Regardless &#8211; thanks Doug for the reminder that we need to think more creatively about how to approach clusters.</p>
<p>Jeff</p>
]]></content:encoded>
	</item>
</channel>
</rss>