Recently I deployed a new Linux-based VM for our Aruba Wireless Network (specifically for AirWave). It’s a CentOS-based installation, but a custom ISO that includes the AirWave installer – you can get it form Aruba’s website.
Anyways, it says that before you install AirWave, you should install VMware Tools. I did the typical “VM > Guest > Install/Upgrade VMware Tools” menu option from the Console window, but nothing happened. I guess I was hoping it would automatically install, since I haven’t done it in Linux (with or without a GUI) before. Because it didn’t auto-install, I took to the web and found this helpful article from a fellow WordPress blogger:
Initiate the VMware tools install on your CentOS 6.2 VM. Open a SSH session to your VM and copy/paste this:
[yum -y install perl] - *step skipped for my install
mount /dev/cdrom /mnt/cdrom
cp /mnt/cdrom/VMwareTools-*.tar.gz /tmp
tar -zxf /tmp/VMwareTools-*.tar.gz -C /tmp
rm -f /tmp/VMwareTools-*.tar.gz
rm -rf /tmp/vmware-tools-distrib
Bam! VMware Tools installed!
During the cleanup from a recent DR test at work, I ran into an issue where I couldn’t migrate guests between different hosts. We have three hosts in a HA/DRS cluster, so if I couldn’t move the guests around there is no way that VMware could automatically do this either.
Not exactly sure why I couldn’t do this, I figured I would reboot each host one at a time. The first host was not able to be put into maintenance mode, since the guest vMotions failed. I also realized that this was the host I was trying to get guests off of when the vMotions failed earlier. Other hosts seemed to vMotion guests between them fine, it was only an issue with “host 1.” I also noticed that I could not open the console for any guest on this particular host.
I checked in the Event Logs in vCenter and found an error similar to (I forgot to document the EXACT error message wording): RAMDISK root is full on host1. I didn’t find much on the Internet regarding this error, other than the actual ESXi OS must have a full partition and I should clean it. I wasn’t quite sure what to delete, but I did find a command to identify what part of the host was full (vdf -h). I called VMware for support and, while helpful, they didn’t show me anything I didn’t already find on my own.
I found the partition that was full was only 32 MB and there was a 30 MB /var/log/hpHelper.log (we have HP hosts) file. I copied the log file using WinSCP, then deleted it from the host, and restarted the host services. After that, vMotions to/from that host were successful and I could put it in maintenance mode to reboot it.
– I couldn’t vMotion guests between hosts at one of our datacenter locations.
– The event log in the vSphere client showed that ramdisk root is full on host1.
– SSHing into host1 and running `vdf -h` shows that ramdisk root has 0% free space out of 32 MB.
– I looked around and found that /var/log was very large.
– I found a 30 MB file in /var/log called hpHelper.log.
– I used WinSCP to copy that file to my desktop and read it, it’s just a log file.
– I ran `rm /var/log/hpHelper.log` and `touch hpHelper.log` to delete and re-create the log file.
– I restarted the management services on the host – `./sbin/services.sh restart`
– Now I can open the console on various guests and I can vMotion between hosts.
My mobile device of choice at work is a 13″ MacBook Air. It’s a great system that is extremely lightweight (see: portable). Unfortunately, I need the ability to console into switches to make configuration changes and this computer does not have a serial port. Thankfully, I have a USB to Serial Adapter that I keep in my “go bag.” It doesn’t work out-of-the-box in OS X 10.8 Mountain Lion, but it isn’t too difficult to configure.
I followed a few articles I found online (primarily, this one: http://plugable.com/2011/07/12/installing-a-usb-serial-adapter-on-mac-os-x).
- Plug in the USB to Serial Adapter.
- Click Apple menu > About This Mac > More Info … > System Report …
- Click on Hardware > USB > verify you see USB-Serial Controller D on the right.
- Download the driver here: http://plugable.com/drivers/prolific (I downloaded the PL2303 MacOSX10.6 dmg v1.4.0.zip file).
- Install the driver, then reboot the computer.
- Click on System Preferences > Network > “+” > Interface. If it was installed properly, you will see something like “USB-Serial Controller D.”
- Click cancel.
Now that the device and driver are installed properly, you can download a terminal emulator like ZTerm to make a connection to the serial device.
Recently, I was in the process of installing a new Windows Server 2008 R2 domain controller into a Server 2003 Native domain with a single Server 2003 domain controller.
During the process, I ran all of the necessary checks (ie: dcdiag, etc.) to verify everything in the domain looked correct. I received an issue when running `netdom query fsmo` that basically said it could not complete the operation. I used the GUI to verify all of the FSMO roles and found that there was an error listed in the schema master section.
Having never come across this before, I did research and found this Microsoft KB (255504) that explained how to seize a FSMO role if one is “lost.”
WARNING: Make sure you are absolutely certain that the FSMO role is lost and the system that used to hold this role is not just having issues or is powered off. If at all possible, it would be better to “fix” the issue, rather than just overwrite the settings with a new FSMO role holder.
Here is the text from the KB that I followed and that successfully let me take over the lost schema role (I performed these steps on the Windows Server 2003 domain controller before promoting the Windows Server 2008 R2 system):
Seize FSMO roles
To seize the FSMO roles by using the Ntdsutil utility, follow these steps:
- Log on to a Windows 2000 Server-based or Windows Server 2003-based member computer or domain controller that is located in the forest where FSMO roles are being seized. We recommend that you log on to the domain controller that you are assigning FSMO roles to. The logged-on user should be a member of the Enterprise Administrators group to transfer schema or domain naming master roles, or a member of the Domain Administrators group of the domain where the PDC emulator, RID master and the Infrastructure master roles are being transferred.
- Click Start, click Run, type ntdsutil in the Open box, and then click OK.
- Type roles, and then press ENTER.
- Type connections, and then press ENTER.
- Type connect to server servername, and then press ENTER, where servername is the name of the domain controller that you want to assign the FSMO role to.
- At the server connections prompt, type q, and then press ENTER.
- Type seize role, where role is the role that you want to seize. For a list of roles that you can seize, type ? at the fsmo maintenance prompt, and then press ENTER, or see the list of roles at the start of this article. For example, to seize the RID master role, type seize rid master. The one exception is for the PDC emulator role, whose syntax is seize pdc, notseize pdc emulator.
- At the fsmo maintenance prompt, type q, and then press ENTER to gain access to the ntdsutil prompt. Type q, and then press ENTER to quit the Ntdsutil utility.
I recently wanted to look at what Group Policy options existed for managing a piece of software, so I downloaded the relevant ADMX files from Microsoft and couldn’t remember what to do with them.
After some quick searching, I found a blog post that answered my question – http://tigermatt.wordpress.com/2009/06/06/where-do-i-put-my-admx-files/
- Copy the ADMX files to %systemroot%\PolicyDefinitions (%systemroot% should be C:\WINDOWS).
- Copy the ADML files (if they exist) to %systemroot%\PolicyDefinitions\en-US (again, %systemroot% should be C:\WINDOWS).
- Restart your Group Policy Management Console for the policy files to appear, no need to add a new template.
Also, a good piece of information that was noted on this blog post was that the ADMX/ADML files don’t need to be present on every machine on the network, just on the machine where you are creating/editing the policies.
We use a cool piece of open source software at work in our labs called iTalc that allows teachers to monitor all of the computers in a given computer lab. This not only lets the teacher keep an eye on the kids and what they’re doing (both to keep them on task and to make sure they’re doing their work correctly), but it also allows them to lock each student PC screen when the teacher wants their attention and allows the teacher to project his/her screen to every student PC or use a student PC as an example and project that screen to every other student. Like I said, it’s a very cool application.
The other day my co-worker asked me to work with him on a new lab setup because he was having issues with iTalc. I wanted to document our findings for the future and for others who might be having similar issues (since we couldn’t find too much about iTalc out there on the Internet – though they do have a wonderful support wiki and EduGeek has good forums).
- iTalc uses VNC as the underlying technology and we use UltraVNC for remote support.
- When UltraVNC and iTalc are installed on the same system, they fight over the port (5900) and UltraVNC seemed to win every time – the iTalc agent would just keep closing and re-opening itself.
- iTalc 2.0.0 can “supposedly” be told to run on another port (same with UltraVNC), but we didn’t have luck getting the iTalc agent to run on a different port.
- We had to uninstall UltraVNC on the lab PCs on the network. Since we deploy UltraVNC via a group policy, we had to deny the application of that policy to the computers in the labs then uninstall UltraVNC.
- We installed the latest iTalc agents, version 2.0.0, on each PC in the lab (they were Windows 7 SP1 64-bit computers).
- The last system I planned to install was the teacher station, or the iTalc master, which also needed to have UltraVNC uninstalled, but is a Windows XP SP3 32-bit computer.
- iTalc 2.0.0 32-bit did not run successfully on this PC.
- After much research and testing, I found others complaining about iTalc 2.0.0 on a 32-bit system. I finally found someone who had a version of iTalc 2.0.0 32-bit that had a different date associated with the file than what was posted on SourceForge.net. Furthermore, this other installation file asks you to install the “Babylon” search engine (which you can say no to). If I ever find where I got this file from, I’ll post a link, since I can’t seem to find it right now.
- iTalc works!
Lessons learned are: don’t run iTalc and UltraVNC on the same computer, upgrade/test the teacher PC and one student PC at the same time (having already done all of the student PCs and not being able to get the teacher PC to work was not fun), keep teacher and student PC software versions consistent (Windows 7 32-bit vs. Windows 64-bit), wait for iTalc version 2.0.1 to come out 🙂
I love service status webpages – there are so many services users rely on every day that are in the cloud and it’s nice to be able to determine whether a particular service is having issues or not.
I’ve known about Google’s status page for a while: http://www.google.com/appsstatus
But more recently I came across a link to Apple’s “cloud services” (aka iCloud-related): http://www.apple.com/support/systemstatus/ I found the website when there was an iMessage outage a month or more ago. Granted I already knew about the iMessage problem, but it was nice to know Apple could confirm an issue.
Microsoft has one for Office 365 too, but you need an account to view it, so I won’t post a link.
I have been using a piece of open source software called mRemote (multi-Remote) for a few years. mRemote is an application that allows users to manage multiple remote desktop connections. While it has many useful features (i.e.: it can connect to SSH, VNC, and RDP from one interface), the main features I like about mRemote are: tabbed remote desktop sessions and the creation of a configuration file to “bookmark” your connections. Unfortunately, the development of mRemote had ceased when the primary developer Felix Deimel started working for a company with a similar (paid) product called Royal TS. Personally, Royal TS sounds amazing (cross platform, syncing of configurations, etc.), but my company won’t pay for it.
When I arrived at my current company, I installed a new fork of the old open source mRemote application called mRemoteNG (multi-Remote Next Generation). I set up the application with all of our servers in a configuration file, but wanted to share with my co-workers, so I wrote the .bat file below to allow my co-workers to share my custom configuration file. This script is very simple and asks for input (u for upload, d for download, and q for quit). You’ll need to modify the script to work in your environment (change the variable for the remote path and confirm the variable for the local path) and you’ll need to make sure you have a Windows share setup that other users can access.
If you’re not planning to share the configuration file, this script could also be used to backup a configuration file – you could change the paths to a personal share or a DropBox folder.
REM Author: Your Name Here
REM Date Updated: Date
REM Purpose: CLI for updating an mRemote configuration file.
REM explain usage
echo u - uploads your local mRemote configuration to the server.
echo d - downloads the server mRemote configuration to your computer.
echo q - quits program.
REM define global variable for server path
REM go to startingpoint function
REM startingpoint function - displays options and directs to other functions based on selection
set /p task="Press (u)pload, (d)ownload, (q)uit: "
IF %task%==u (goto upload)
IF %task%==d (goto download)
IF %task%==q (goto done)
IF NOT %task%==d IF NOT %task%==u (goto notvalid)
REM download function - downloads the configuration file from the server to the local hard drive (and creates local backup)
echo Creating backup and downloading file ...
copy /V /Y %localpath% %localpath%_BACKUP
copy /V /Y %serverpath% %localpath%
REM upload function - uploads the configuration file from the local hard drive to the server (and creates remote backup)
echo Creating backup and uploading file ...
copy /V /Y %serverpath% %serverpath%_BACKUP
copy /V /Y %localpath% %serverpath%
REM notvalid function - if a user does not press u, d, or q, function explains usage again and takes back to starting point
echo Not a valid choice ...
echo Acceptable choices are:
echo u - upload
echo d - download
echo q - quit
echo ... Please try again.
REM done function - exits program
I’ve been struggling with my personal backups for a while. I use a MacBook (white 13″ mid-2007, OS X 10.7.5 Lion) at home as my primary computer and have always used Time Machine. How I use Time Machine has changed over the last few years:
- I started using Time Machine with a locally connected external portable USB Western Digital hard drive. This always worked well, but required me to remember to plug in the external HDD to run a backup.
- I bought an AirPort Extreme Base Station (AEBS) for Christmas 2011 and connected my USB printer, my wife’s backup HDD, and my backup HDD via a powered USB hub. In doing this, I was able to share out each hard drive independently and re-seed our Time Machine backups. This worked great and automatically backed up our data hourly – no physical connections required!
- I found a Good Friday deal this past year (2012) and bought a 1-year subscription to CrashPlan+ for 98% off (something like $2.40 for unlimited data and up to 10 computers for the year). I always liked Time Machine, but figured I should have an off-site backup too and for under $5 for the year I couldn’t pass it up.
My local wireless Time Machine backup combined with my offsite wireless CrashPlan backup seemed like the ultimate solution, and it was for about 10 days.
Starting about a week before Christmas, my Time Machine backup stopped working. After much troubleshooting, I realized that I couldn’t even connect to the AirDisk shares on my AEBS. Rebooting the base station fixed this problem temporarily. Reformatting my HDD and re-seeding my backup only worked as long as it took to finish the initial backup, then continued to fail. Again, I couldn’t connect to the AirDisk shares without rebooting the base station. I did more research and tested more theories, but Time Machine would never continue to run for long. I finally shut off Time Machine and left it off for a few days. During this time, I could connect to the AirDisk shares anytime I tried. Once enabling Time Machine again, it ran for a few backups, then failed again.
Eventually it dawned on me that Time Machine stopped working around the time I started using CrashPlan+. I disabled CrashPlan+ and Time Machine has been running successfully for about 4 days now, no hiccups whatsoever!
It seems that much like having more than 1 antivirus solution running on the same system isn’t good/recommended, multiple backup solutions running on the same system also doesn’t appear to be the best idea. Unfortunate too, CrashPlan+ has an article about how their product compliments Time Machine. Personally, I will rely more on Time Machine since I have it set to backup everything on my hard drive. I only used CrashPlan+ to backup certain areas of my hard drive that I wanted offsite. So, If i can keep Time Machine working I’ll be happy.
Some additional ideas I wrote down for troubleshooting that I haven’t gotten to try yet include:
- Wipe AEBS settings/configuration – I already downgraded firmware to 7.6 and then re-applied 7.6.1 when the downgrade didn’t work.
- Test wireless Time Machine backup without USB hub and enable CrashPlan+ again.
- Test directly connecting my portable HDD to my laptop and running the Time Machine backup again with CrashPlan+ enabled.
If I ever find out exactly why the two technologies don’t work together or if I ever discover a workaround, I’ll post again. One theory I had was that there could be an issue if they’re trying to backup the same file at the same time. I couldn’t find a way to schedule either (for example, running them every other hour), though I’ll admit I didn’t look too hard yet.
At work I need to use a custom DNS search suffix list on my primary network adapter so that I can resolve hostnames on my primary Active Directory (AD) domain and so that I can resolve hostnames against a secondary AD domain that is also on my network.
While I have been running with a custom search suffix list for a few months now, I just found a small problem with the setup: when using the nslookup command to lookup up an IP address against an external source, the last DNS search suffix in the list was appended to everything I looked up. For example, if the suffixes in my list were a.org and b.net, my commands looked like this:
Default Server: <local DNS server>
Address: <local DNS server address>
> server 184.108.40.206
Default Server: b.resolvers.Level3.net
*** b.resolvers.Level3.net can't find google.com.b.net : Non-existant domain
I had no idea why the .b.net kept appending itself to the results. Eventually, after doing some research, I found that this was “normal” with the nslookup command in Windows and in order to get around it, I’d need to add a period (.) to the end of the request. For example:
Default Server: <local DNS server>
Address: <local DNS server address>
> server 220.127.116.11
Default Server: b.resolvers.Level3.net
Mystery solved! I can use nslookup again to resolve against an external DNS server if I end the request with a period. Also to note, that technically a fully-qualified domain name (FQDN) ends in a period anyways – it’s just usually assumed.