Grumbles
Probably going to charge 200% markup on the cost of the UK version...
5267 publicly visible posts • joined 4 Dec 2007
If the law cannot be explained in terms that those subject to that law can understand, you need to ditch the law.
It is utterly inappropriate to suggest that we should be governed by laws that we have no hope of understanding.
If there was a problem with an inaccurate understanding of the terms, time should have been set aside to explain clearly what the concepts mean and there should have been some concept checking to make sure the jargon is understood. The meanings should have been gone over again before the jury retired. The judge should have ensured this and should have noted use of jargon and pulled up the lawyers when they used it and made sure both sides agreed on the definitions - perhaps keeping a "cheat-sheet" available to the jury. We don't need to switch to an inquisitorial system for this, this is what the judge is for - to make sure things are conducted properly.
There will be times when justice is not well-served by jury trial and a "better" outcome could have been achieved some other way, but a jury trial is the mechanism by which the people keep control of the legal system. That is far more important than this one trial.
More food?
Do we really need more corn? Does India, the land of lentils, need protein-enhanced potatoes? If you live in an arid area, how long will a strategy of planting crops which can suck more moisture out of the ground, last? Has anyone else noticed the rise in problems with mass-consumed products? Gluten, dairy, sugar and increasingly, soy?
Everyone I've met who is interested in food says that your best food is the stuff people haven't processed, that is left on the plant until ripe and then eaten.
GM is all about factory efficiencies. Yes we can produce "more" food, but is it just swollen with water or has the nutrient content per lb/kg been preserved? Yes people want red tomatoes, but that's because historically red tomatoes mean ripe tomatoes. If you make them red when they are unripe you can store/sell them for longer, but that is just tricking the customers, not making things better. Yes we can make the crops insecticide-resistant and then kill all the insects, but that will kill all the birds and hence other types of plants which we do actually need to have around. All these are costs are not incurred by the GM companies.
In the UK organic food is governed by the Soil Association. If you care for the soil and the plants thrive and you don't have to go around "altering" out large parts of the ecosystem. Efficiency should not always be the highest priority.
GM might be efficient in the short-term, but so is slash and burn. That doesn't make it the right thing to do.
Why would a touch GUI transform it into a success?
And why would Valve be that worried? It hasn't felt the need to get into the tablet/phone game market - all its catalog is on large devices. Casual touch games are almost too cheap to make money on when they are new - the only way to make money is be selling millions and if you don't own the platform, you don't get a chance to do that.
Apple's store doesn't really compete (effectively) with Steam and that's been on the Mac for a while, though MS is likely to be more aggressive in its pricing.
MS could do well in mobile gaming, but I suspect its Nintendo and PS who need to be worried about that. I can't see keyboard/mouse gamers having more than a passing interest in touchscreen entertainment - its even more inaccurate than a console.
Oz retailers charge whatever they think the market will bear. My local woolworths supermarket is charging 10c/litre more than the bp station a couple of hundred yards away on the other side of the road and aud 2 more for a bathroom scales battery than the corner shop/pharmacy at the top of my road. Tomatos are almost AUD12/kg
Promiscuous mode is the least of your problems. You company has installed a cert into your PC which makes it look as though you have an SSL connection to your bank, when it really only goes as far as your proxy, where everything is decrypted (in a misguided attempt at data loss prevention) and then re-encrypted.
If you need maximum TPS, ARM probably isn't your thing.
The tech is also quite new so economies of scale haven't really kicked in yet. At this stage its a question of selling tech to get people skilled up.
I'd like to see some costings when you compare lots of ARM CPUs to x86 hardware + virtualisation costs.
If you are using hierarchical storage, then you get benefit from having fewer, but high-capacity low-level storage. Fewer disks (hopefully) mean fewer failures (maintenance callout costs) and more cash to spend on cache further up the hierarchy. Your upper tiers take the IOPS strain for you.
Plus, there are plenty of applications where you want lots of data, but don't need to access it very often. Say hello to the home market. Anyone doing photography might appreciate it. Lots of very hi-res images, but you only need to access a few at a time. Or backups. Yes I want timemachine to go back months, but I rarely need access to that data. Or I might be spooling off backups to disk and then transferring them to tape.
Requirements are generally tuned for what is available. There usually is a niche for "more", even if it isn't in the standard server market.
It does seem like a rather "special" application. They cost about 3/4 of an old core2 but do so much less and are likely to be on a hobbled motherboard.
Why would you stick a brand new, bloated OS on a low-power cpu? Are consumers going to be able to get a cheap server version of W8 with the desktop removed from the kernel?
Product positioning gone bad, I think.
> Its not done by default in linux, just like it isnt in windows.
You must have been using a numpty distro. IME it is normal in linux distros, but it is a distro rather than kernel option.
It's been a few years since I've tried, but windows used to make a hash of it, since drive space was calculated at the "drive letter" level - directory mounted volumes appeared to be a kludge.
Also, separation is not as useful under windows, since the UIDs are unique to the installation and you'll end up with "unknown user" which isn't easy to fix (unless you've got central authentication running from somewhere else) since user data appears to be scattered all over the system - file and registry.
For example, I keep a spare partition so I can do a clean install to a different partition for a major upgrade and just remount the user directories. I made the mistake of trying to do the same thing for a windows install. I suspect there is a migration agent somewhere which will do it cleanly, but linux... well it's just less hassle. I also tried changing the "special folder" location - that didn't go well.
I'm sure all this stuff can be done under windows, it just seems like more trouble than its worth.
How true. You can change all sorts of things in kmail (which I love) but then you get things like text almost flat up against a border, which looks terrible. It looks as though underlying data structures have made it into the gui with grouping headers, boxes and lines around gui elements, and so on. It would probably be better to have fewer features and better looks. That's no bad reflection on the devs, but we do need some more influential aesthetic designers.
The gnome 3 gui may put off linux devs and tifkam may be a flop, but in the end its
about the apps. Put ms office on the desktop and the year of the linux desktop in business would be here. It isn't the features, its about risk and confidence.
True, but valve may not be going for AAA. A simple, cheap console could be very appealing and if its linux based, it could offer very cool facilities. Amazon integration as good as a kindle, for content acquisition? hooks into the home stereo (likely to be near the tv), how about a calibre ebook server, general NAS or the obvious mythtv? If Valve can make a linux box really useful and there are lots of them around, then we might see more games coming onboard. There are plenty of basic linux games which Valve could give away or integrate into its cloud.
Does anyone else think that if the government thinks suicide sites need shutting down then they *should* be regulating them?
Its very very important! So important, we're not going to devote any time to it. We're just going to throw random threats around and hope someone else does it for us.
If the government wants something regulated, it should jolly well regulate it, not try to slip in restrictions by the back door. Go ahead, say you want to censor it and provide the justification.
It isn't efficient. UDP was designed for applications where lossiness isn't an issue, whereas TCP adds processing overheads in sequence numbers and bandwidth for acknowledgements and latency when packets are lost, with TCP windows.
As long as your data/application can survive data loss then UDP should be fine.
However, http allows you to cross proxies and firewalls to get into corporates whereas new ports would never be allowed.
I think realplayer fouled things up for everyone by being such a pig of a protocol/implementation. Many mobile phones required realplayer to have fixed source & destination ports which made NAT mostly impractical, even when the desktop version of the protocol had moved on and allowed dynamic source ports.
HTTP is not so much an application as a transport these days, with data being held in proprietary apps on each end.
It is mostly about content protection - the antithesis of HTML5 and the web which seeks to expose data and make it usable everywhere. Any DRM will have to involve binary blobs and inaccessible data because it is almost guaranteed that the system will be broken and the "standard" will need to be changed.
The idea appears to be to allow a stream but not a download, desperately trying to shove the digital genie back in the bottle and revert to pure analogue broadcast with no VCR. If you use download rather than streaming, the user merely waits a bit and then picks everything up at local network or media speeds.
Having said that, there are occasions when high demand might make degrading quality useful. Streaming the FA cup final might be one of them - you don't want to hit a "pause" as the ball heads towards goal. However, for that application we have a wonderful broadcast medium called "telly" which works quite well. Degrading quality is for when you haven't built the network you should have.
The problem is that these things are expensive and with expense comes risk-aversion - so fewer purchases for the non-market leader.
Then there is the problem that cheaper tablets self-select to those with less money who are more likely to go for a laptop instead. And also there is a bit of cheapness in the build quality which Apple wouldn't countenance - so they don't look as good either.
I want an x86/arm laptop hybrid with detachable screen. A tablet is too limited to spend that much on a good screen and battery.
I haven't looked closely and I may be completely wrong, but if you use a wireless device, you can improve things with a local proxy (assuming http streaming). With a wireless device, you are most likely to need retransmits (++latency) in the last few meters of network. A local proxy *may* prevent the need to go back out to the internet for content.
I only tried this once and it could have been a fluke that when I switched to the proxy things got better, other commenters please chime in. Does squid do anything clever to improve things with streaming? Could it be made to pre-cache content and ask for the entire file as fast as possible, even if only a small bit of it has been requested?
But downloads via my HDHR don't come off my internet allowance.
If you treat myth as an appliance, you don't really need to fiddle with it much. Its the desire to constantly upgrade which will get you into trouble. If you like to do that sort of thing, I recommend using a VM which excludes the main filestore but which includes the database. Then you can snapshot and rollback very easily.
Just set up something which monitors the mythbackend daemon and restarts it if required and you'll be fine. A couple of versions ago I had issues with stability, but recently it seems all fine.
But I do have to agree. The networks in Oz seem to think you won't come back if they go to an ad-break without a cliffhanger. That usually spoils the dramatic tension completely at the worst times - pirate content is just a far better experience.
Well you could treat them the same and just develop TIFKAM apps, but no-one would buy them for the desktop. You could develop a cool app which needs gig-ethernet data speeds, but no-one would buy it for phones or tablets.
Its just PR drivel. Actually, its worse for MS. By setting up the expectation that tablet, phone and desktop are all the same, they are setting up both devs and customers for a massive disappointment. Got IE on a tablet? Great, but why does the website say I need to install something called "flash"? I transferred my 300 page document from my work pc to my tablet, but I find it quite slow scrolling down, why is that? Why can't I find my colleagues' email addresses when I'm at home on my phone?
If Apple did one thing that no-one else had done before, it was to completely change the expectations of users, by not having a physical keyboard or mouse or anything which looked like OSX on the ip*d. Managing expectations is key to a happy customer and Ballmer's doing it all wrong.
but have they sorted out the mess that causes with checking disk sizes?
50Gig: c:\
1000Gig c:\mybigdisk
I want to install a new program in c:\mybigdisk and the installer checks the disk size of c: and doesn't allow it.
I haven't tried it in a while, so it may have been fixed - answers on a postcard.
That said, I'm really annoyed at KDE, the way it holds volumes mounts in the GUI layer rather than updating the underlying OS. At least OSX bungs everything in /Volumes so the whole OS has access and not just the GUI.