Isn't MS Script Editor essentially VBScript?
I didn'r read the utf16 article, but basically utf8 is really slick, because it can do anything. But it stops being efficient if you represent one string with lots of different language characters in it.
Although I find it hard to understand why people complain about inefficient utf-8 (as in you may burn 4 bytes per character for the whole document in some cases) but they don't complain about the rest of the fucked up shit in the computer industry like.. oh... windows vista for example.
Anyway, so really you only need utf8, but if you're going to do lots of 4 byte characters, you're better off with utf 16 which makes more efficient use of the space for the same thing.
but if... like odf does... a gzip runs over that text it will compress better, and give you back your bigger space.
I should think UTF16 compresses at least as well as UTF8, but I haven't actually done any benchmarks to make sure.
UTF8 is pretty smart, though. It's truly a shame Microsoft went with UTF16.
If it's old old, remember there is a 120gig barrier for standard IDE, for old BIOSes and operating systems.
a good collection.
*prints and puts under the cussion*
Well I didn't want to go to our Tech department because it's outsourced adn every request costs money. I thought I could figure it out but the IT Director (who is not an IT person really) overheard me talking to my boss about what I wanted to do and INSISTED that I submit a Sysaid and let them handle it. She said my time was better spent and I'm about the only person in the agency who would come up with a juicy request like that.
I may still try to figure out scripting stuff on my own... just cause I keep needing to use it.
So, does anyone have any suggestions for me on laptop hard disk brands??
So if UTF-8 essentially burns several bytes per character -
1. Does it use more bytes per character for normal English?
2. Does it use the same for, say, Chinese?
... In other words, would a document with 1000 English characters be smaller than a document with 1000 Chinese characters?
(And do IP packets still get addressed with standard ASCII, or can they use UTF-8 for that too somehow?)
the chinese will definitely be bigger. how big depends on how high in the unicode namespace the chinese chars are...
you pro'lly mean DNS by "ip" right? since ips are just numbers, they don't have a relation to UTF8.
DNS does something similar as you might have seen in your emails with base64 with umlaut domains... so a purely chinese domain might take up to 5 times what a generic english domain would take in glyphs.
Part of the brilliance of UTF-8 is that the representation of ASCII characters 0-127 does not change at all. It's strictly one byte per character, and characters 0-127 of UTF-8 are identical to characters 0-127 of ASCII. Nothing changes.
It's only when you get into the international character sets (such as Chinese) that you start to insert multibyte characters into the data stream. When you have a character in the range of 128-255, it means that you're constructing a multibyte character in UTF-8. This allows all characters to be representable using (mostly) existing software, without requiring wasted space for representation of ASCII characters.
Hmm, actually it's 1-127 and 128-254. 0 is still a null terminator, and 255 is used as an optional "byte order mark" (U+FEFF). The BOM is normally only used in UTF-16, but some software naively retains it when downconverting to UTF-8.
When you asked 'does it use more bytes per character for normal English', the answer is 'no'.
When you asked 'does it use the same for, say, Chinese', the answer is 'ASCII doesn't have a way to represent Chinese'. But that's a pedantic answer.
I'm not 100% sure, but I think it's possible to use more characters to represent some languages in UTF-8 than UTF-16 (the Windows solution). But for English, UTF-8 wins over UTF-16.
Correct. For typical English and Western European languages, UTF-8 is a more compact encoding. However, for "CJK" Chinese/Japanese/Korean, UTF-16 is the more compact encoding. All of those characters are from a codepoint range that encodes to three UTF-8 bytes or (usually) two UTF-16 bytes.
the drive that died was a western digital but thats a 3.5 desktop drive.
so much for SMART warning me of impending failure...
A stupid question...did you have some SMART tool software enabled? As far as I know, no OS has these tools installed by default.
So I was about to format the drive when i decided to run recovery boot.
Looked at the boot drive it was fine.
Then I rmemberd something about the slave drive being the boot drive, so I told my bios to boot from the slave, and VOILA. working machine.
So I'm not sure what went wrong.
I'm running chkdsk on all the drives, I have a feeling one of my drives is on the way out, but somebody (BIOS or MBR) lost track of which drive was the boot drive.
I still think something's wrong, but now I don't know what.
I could just be that it shutdown due to overheating. I'll never know.
but there was definetly a boot or two where the bios had a hard time detecting IDE devices.
I'm sure something will go wrong at the worst time and I won't be able to figure out what it is.
but at least now I have the opoortunity to back everything upl.