No announcement yet.

Help Implementing Vuforia Code in Android Studio

  • Filter
  • Time
  • Show
Clear All
new posts

  • Help Implementing Vuforia Code in Android Studio

    After getting the ConceptVuforiaNavigation Opmode to run on our Moto G4 phones, and properly identify the four Rover Ruckus targets, our team decided to try the ConceptVuforiaNavigationWebcam Opmode (in hopes of using a higher end web cam for navigation, as well as identifying the Gold Mineral in Auton), and our students are completely stuck with multiple errors in Android Studio stopping them from compiling.

    Our code is below, and we've attached screen caps of the errors we're encountering, and here are the details on our setup from the Philbot list:

    1) What type of robot hardware are you using? Modern Robotics, REV Expansion or mixed.

    Single REV expansion hub with updated firmware

    2) What type of phone are you using? ZTE, Moto G2, Motor 4 Play or other.

    Moto G4 Play

    3) How did you transfer the FTC App to your Phone? Android Studio Download, Play Store download, Other?

    Android Studio Download Via Wireless ADB

    4) CRITICAL !!!! If you used Android Studio to deploy your app... DID YOU TURN OFF AUTO RUN? You MUST do this. !!!!!!!!!!!!!!!!!!!!!!!
    File / Settings / Build-Execution-Deployment / Instant Run (uncheck)


    5) What is the version of the RC App? Use the menu to show the About screen. Rev 3.4 is currently rev, but it could be 3.1 or 2.62 or even 2.4 if it's from early last year.


    6) If you are having trouble configuring your robot, are you using the RC phone to do it?
    If not (you are using the DS phone), you MUST have paired with the robot and be connected (Numbers showing in Ping Times)

    We haven't gotten this far with this particular Opmode issue, however, the robot operates 100% in other Teleop Modes.

    7) Is this a problem that has started since you made a change to the system? What was that change?

    No, the particular Opmode (ConceptVuforiaNavigationWebcam) has never worked

    8) If you need help doing something new, have you checked the sample programs to see if there is already an example for you to emulate?

    We are using the sample program

    9) If you are having crashes or disconnects or anomalous behavior, be ready to provide a copy of you log file (found on phone) to help in debugging.

    N/A (as far as we know)

    10) New Vuforia Key (Just in case)

    /* Copyright (c) 2017 FIRST. All rights reserved.
     * Redistribution and use in source and binary forms, with or without modification,
     * are permitted (subject to the limitations in the disclaimer below) provided that
     * the following conditions are met:
     * Redistributions of source code must retain the above copyright notice, this list
     * of conditions and the following disclaimer.
     * Redistributions in binary form must reproduce the above copyright notice, this
     * list of conditions and the following disclaimer in the documentation and/or
     * other materials provided with the distribution.
     * Neither the name of FIRST nor the names of its contributors may be used to endorse or
     * promote products derived from this software without specific prior written permission.
    package org.firstinspires.ftc.robotcontroller.external.samples;
    import com.qualcomm.robotcore.eventloop.opmode.TeleOp;
    import com.qualcomm.robotcore.eventloop.opmode.Disabled;
    import com.qualcomm.robotcore.eventloop.opmode.LinearOpMode;
    import com.qualcomm.robotcore.util.RobotLog;
    import com.qualcomm.robotcore.util.ThreadPool;
    import com.vuforia.Frame;
    import org.firstinspires.ftc.robotcore.external.ClassFactory;
    import org.firstinspires.ftc.robotcore.external.hardware.Consumer;
    import org.firstinspires.ftc.robotcore.external.hardware.Continuation;
    import org.firstinspires.ftc.robotcore.external.matrices.MatrixF;
    import org.firstinspires.ftc.robotcore.external.matrices.OpenGLMatrix;
    import org.firstinspires.ftc.robotcore.external.navigation.AngleUnit;
    import org.firstinspires.ftc.robotcore.external.navigation.AxesOrder;
    import org.firstinspires.ftc.robotcore.external.navigation.AxesReference;
    import org.firstinspires.ftc.robotcore.external.navigation.Orientation;
    import org.firstinspires.ftc.robotcore.external.navigation.VuforiaLocalizer;
    import org.firstinspires.ftc.robotcore.external.navigation.VuforiaTrackable;
    import org.firstinspires.ftc.robotcore.external.navigation.VuforiaTrackableDefaultListener;
    import org.firstinspires.ftc.robotcore.external.navigation.VuforiaTrackables;
    import org.firstinspires.ftc.robotcore.internal.system.AppUtil;
    import java.util.ArrayList;
    import java.util.List;
    import java.util.Locale;
     * This 2016-2017 OpMode illustrates the basics of using the Vuforia localizer to determine
     * positioning and orientation of robot on the FTC field.
     * The code is structured as a LinearOpMode
     * Vuforia uses the phone's camera to inspect it's surroundings, and attempt to locate target images.
     * When images are located, Vuforia is able to determine the position and orientation of the
     * image relative to the camera.  This sample code than combines that information with a
     * knowledge of where the target images are on the field, to determine the location of the camera.
     * This example assumes a "diamond" field configuration where the red and blue alliance stations
     * are adjacent on the corner of the field furthest from the audience.
     * From the Audience perspective, the Red driver station is on the right.
     * The two vision target are located on the two walls closest to the audience, facing in.
     * The Stones are on the RED side of the field, and the Chips are on the Blue side.
     * A final calculation then uses the location of the camera on the robot to determine the
     * robot's location and orientation on the field.
     * @see VuforiaLocalizer
     * @see VuforiaTrackableDefaultListener
     * see  ftc_app/doc/tutorial/FTC_FieldCoordinateSystemDefinition.pdf
     * Use Android Studio to Copy this Class, and Paste it into your team's code folder with a new name.
     * Remove or comment out the @Disabled line to add this opmode to the Driver Station OpMode list.
     * IMPORTANT: In order to use this OpMode, you need to obtain your own Vuforia license key as
     * is explained below.
    @TeleOp(name="Concept: Vuforia Nav Webcam", group ="Concept")
    public class ConceptVuforiaNavigationWebcam extends LinearOpMode {
        public static final String TAG = "Vuforia Navigation Sample";
        OpenGLMatrix lastLocation = null;
         * @see #captureFrameToFile()
        int captureCounter = 0;
        File captureDirectory = AppUtil.ROBOT_DATA_DIR;
         * {@link #vuforia} is the variable we will use to store our instance of the Vuforia
         * localization engine.
        VuforiaLocalizer vuforia;
         * This is the webcam we are to use. As with other hardware devices such as motors and
         * servos, this device is identified using the robot configuration tool in the FTC application.
        WebcamName webcamName;
        public void runOpMode() {
             * Retrieve the camera we are to use.
            webcamName = hardwareMap.get(WebcamName.class, "Webcam 1");
             * To start up Vuforia, tell it the view that we wish to use for camera monitor (on the RC phone);
             * If no camera monitor is desired, use the parameterless constructor instead (commented out below).
            int cameraMonitorViewId = hardwareMap.appContext.getResources().getIdentifier("cameraMonitorViewId", "id", hardwareMap.appContext.getPackageName());
            VuforiaLocalizer.Parameters parameters = new VuforiaLocalizer.Parameters(cameraMonitorViewId);
            // OR...  Do Not Activate the Camera Monitor View, to save power
            // VuforiaLocalizer.Parameters parameters = new VuforiaLocalizer.Parameters();
             * IMPORTANT: You need to obtain your own license key to use Vuforia. The string below with which
             * 'parameters.vuforiaLicenseKey' is initialized is for illustration only, and will not function.
             * A Vuforia 'Development' license key, can be obtained free of charge from the Vuforia developer
             * web site at
             * Vuforia license keys are always 380 characters long, and look as if they contain mostly
             * random data. As an example, here is a example of a fragment of a valid key:
             *      ... yIgIzTqZ4mWjk9wd3cZO9T1axEqzuhxoGlfOOI2dRzKS4T0hQ8kT ...
             * Once you've obtained a license key, copy the string from the Vuforia web site
             * and paste it in to your code on the next line, between the double quotes.
            parameters.vuforiaLicenseKey = "AaxEeJj/////AAABmdLSaHu3GEP4oER0Z4wPyoYCDrpQOx7X1WZ792YNzZwWGHsLEoTDlfqsXqFSF65DRwdrvXt8RCy1oOpTCReV8Mb7NJETFJupTeuqAt9KHsxxSszvgNq6nY4yzLCMWHPWmmh+iKRyJGSKXv4rZ3Z9t1wcPkSZ1p0jBlFX9v4wqGQSumKTKmgpV+133yONI/3EX8UvIJJQMMW65V1SiNaq8xw2NRZMXYU2b8BemEUQTYDmdD1mOoOw5gBSYuNhaWR3JpZ0tGJ2n4CrSQiD+UeGLC1NkA+lFp3XgoYIbHb6p08rdRsvbytOTQ/13fCfiVtpgibzARDbRQQ9PcdTBMphdHjf2ZZmsDG4iDYV2hmUGbMQ";
             * We also indicate which camera on the RC we wish to use.
            parameters.cameraName = webcamName;
             * Instantiate the Vuforia engine
            vuforia = ClassFactory.getInstance().createVuforia(parameters);
             * Because this opmode processes frames in order to write them to a file, we tell Vuforia
             * that we want to ensure that certain frame formats are available in the {@link Frame}s we
             * see.
            /** @see #captureFrameToFile() */
             * Load the data sets that for the trackable objects we wish to track. These particular data
             * sets are stored in the 'assets' part of our application (you'll see them in the Android
             * Studio 'Project' view over there on the left of the screen). You can make your own datasets
             * with the Vuforia Target Manager: PDFs for the
             * example "StonesAndChips", datasets can be found in in this project in the
             * documentation directory.
            VuforiaTrackables stonesAndChips = vuforia.loadTrackablesFromAsset("StonesAndChips");
            VuforiaTrackable redTarget = stonesAndChips.get(0);
            redTarget.setName("RedTarget");  // Stones
            VuforiaTrackable blueTarget = stonesAndChips.get(1);
            blueTarget.setName("BlueTarget");  // Chips
            /** For convenience, gather together all the trackable objects in one easily-iterable collection */
            List<VuforiaTrackable> allTrackables = new ArrayList<VuforiaTrackable>();
             * We use units of mm here because that's the recommended units of measurement for the
             * size values specified in the XML for the ImageTarget trackables in data sets. E.g.:
             *      <ImageTarget name="stones" size="247 173"/>
             * You don't *have to* use mm here, but the units here and the units used in the XML
             * target configuration files *must* correspond for the math to work out correctly.
            float mmPerInch = 25.4f;
            float mmBotWidth = 18 * mmPerInch;            // ... or whatever is right for your robot
            float mmFTCFieldWidth = (12 * 12 - 2) * mmPerInch;   // the FTC field is ~11'10" center-to-center of the glass panels
             * In order for localization to work, we need to tell the system where each target we
             * wish to use for navigation resides on the field, and we need to specify where on the robot
             * the camera resides. These specifications are in the form of <em>transformation matrices.</em>
             * Transformation matrices are a central, important concept in the math here involved in localization.
             * See <a href="">Transformation Matrix</a>
             * for detailed information. Commonly, you'll encounter transformation matrices as instances
             * of the {@link OpenGLMatrix} class.
             * For the most part, you don't need to understand the details of the math of how transformation
             * matrices work inside (as fascinating as that is, truly). Just remember these key points:
             * <ol>
             *     <li>You can put two transformations together to produce a third that combines the effect of
             *     both of them. If, for example, you have a rotation transform R and a translation transform T,
             *     then the combined transformation matrix RT which does the rotation first and then the translation
             *     is given by {@code RT = T.multiplied(R)}. That is, the transforms are multiplied in the
             *     <em>reverse</em> of the chronological order in which they applied.</li>
             *     <li>A common way to create useful transforms is to use methods in the {@link OpenGLMatrix}
             *     class and the Orientation class. See, for example, {@link OpenGLMatrix#translation(float,
             *     float, float)}, {@link OpenGLMatrix#rotation(AngleUnit, float, float, float, float)}, and
             *     {@link Orientation#getRotationMatrix(AxesReference, AxesOrder, AngleUnit, float, float, float)}.
             *     Related methods in {@link OpenGLMatrix}, such as {@link OpenGLMatrix#rotated(AngleUnit,
             *     float, float, float, float)}, are syntactic shorthands for creating a new transform and
             *     then immediately multiplying the receiver by it, which can be convenient at times.</li>
             *     <li>If you want to break open the black box of a transformation matrix to understand
             *     what it's doing inside, use {@link MatrixF#getTranslation()} to fetch how much the
             *     transform will move you in x, y, and z, and use {@link Orientation#getOrientation(MatrixF,
             *     AxesReference, AxesOrder, AngleUnit)} to determine the rotational motion that the transform
             *     will impart. See {@link #format(OpenGLMatrix)} below for an example.</li>
             * </ol>
             * This example places the "stones" image on the perimeter wall to the Left
             *  of the Red Driver station wall.  Similar to the Red Beacon Location on the Res-Q
             * This example places the "chips" image on the perimeter wall to the Right
             *  of the Blue Driver station.  Similar to the Blue Beacon Location on the Res-Q
             * See the doc folder of this project for a description of the Field Coordinate System
             * conventions.
             * Initially the target is conceptually lying at the origin of the Field Coordinate System
             * (the center of the field), facing up.
             * In this configuration, the target's coordinate system aligns with that of the field.
             * In a real situation we'd also account for the vertical (Z) offset of the target,
             * but for simplicity, we ignore that here; for a real robot, you'll want to fix that.
             * To place the Stones Target on the Red Audience wall:
             * - First we rotate it 90 around the field's X axis to flip it upright
             * - Then we rotate it  90 around the field's Z access to face it away from the audience.
             * - Finally, we translate it back along the X axis towards the red audience wall.
            OpenGLMatrix redTargetLocationOnField = OpenGLMatrix
                    /* Then we translate the target off to the RED WALL. Our translation here
                    is a negative translation in X.*/
                    .translation(-mmFTCFieldWidth / 2, 0, 0)
                            /* First, in the fixed (field) coordinate system, we rotate 90deg in X, then 90 in Z */
                            AxesReference.EXTRINSIC, AxesOrder.XZX,
                            AngleUnit.DEGREES, 90, 90, 0));
            RobotLog.ii(TAG, "Red Target=%s", format(redTargetLocationOnField));
             * To place the Stones Target on the Blue Audience wall:
             * - First we rotate it 90 around the field's X axis to flip it upright
             * - Finally, we translate it along the Y axis towards the blue audience wall.
            OpenGLMatrix blueTargetLocationOnField = OpenGLMatrix
                    /* Then we translate the target off to the Blue Audience wall.
                    Our translation here is a positive translation in Y.*/
                    .translation(0, mmFTCFieldWidth / 2, 0)
                            /* First, in the fixed (field) coordinate system, we rotate 90deg in X */
                            AxesReference.EXTRINSIC, AxesOrder.XZX,
                            AngleUnit.DEGREES, 90, 0, 0));
            RobotLog.ii(TAG, "Blue Target=%s", format(blueTargetLocationOnField));
             * We also need to tell Vuforia where the <em>cameras</em> are relative to the robot.
             * Just as there is a Field Coordinate System, so too there is a Robot Coordinate System.
             * The two share many similarities. The origin of the Robot Coordinate System is wherever
             * you choose to make it on the robot, but typically you'd choose somewhere in the middle
             * of the robot. From that origin, the Y axis is horizontal and positive out towards the
             * "front" of the robot (however you choose "front" to be defined), the X axis is horizontal
             * and positive out towards the "right" of the robot (i.e.: 90deg horizontally clockwise from
             * the positive Y axis), and the Z axis is vertical towards the sky.
             * Similarly, for each camera there is a Camera Coordinate System. The origin of a Camera
             * Coordinate System lies in the middle of the sensor inside of the camera. The Z axis is
             * positive coming out of the lens of the camera in a direction perpendicular to the plane
             * of the sensor. When looking at the face of the lens of the camera (down the positive Z
             * axis), the X axis is positive off to the right in the plane of the sensor, and the Y axis
             * is positive out the top of the lens in the plane of the sensor at 90 horizontally
             * counter clockwise from the X axis.
             * Next, there is Phone Coordinate System (for robots that have phones, of course), though
             * with the advent of Vuforia support for Webcams, this coordinate system is less significant
             * than it was previously. The Phone Coordinate System is defined thusly: with the phone in
             * flat front of you in portrait mode (i.e. as it is when running the robot controller app)
             * and you are staring straight at the face of the phone,
             *     * X is positive heading off to your right,
             *     * Y is positive heading up through the top edge of the phone, and
             *     * Z is pointing out of the screen, toward you.
             * The origin of the Phone Coordinate System is at the origin of the Camera Coordinate System
             * of the front-facing camera on the phone.
             * Finally, it is worth noting that trackable Vuforia Image Targets have their <em>own</em>
             * coordinate system (see {@link VuforiaTrackable}. This is sometimes referred to as the
             * Target Coordinate System. In keeping with the above, when looking at the target in its
             * natural orientation, in the Target Coodinate System
             *     * X is positive heading off to your right,
             *     * Y is positive heading up through the top edge of the target, and
             *     * Z is pointing out of the target, toward you.
             * One can observe that the Camera Coordinate System of the front-facing camera on a phone
             * coincides with the Phone Coordinate System. Further, when a phone is placed on its back
             * at the origin of the Robot Coordinate System and aligned appropriately, those coordinate
             * systems also coincide with the Robot Coordinate System. Got it?
             * In this example here, we're going to assume that we put the camera on the right side
             * of the robot (facing outwards, of course). To determine the transformation matrix that
             * describes that location, first consider the camera as lying on its back at the origin
             * of the Robot Coordinate System such that the Camera Coordinate System and Robot Coordinate
             * System coincide. Then the transformation we need is
             *      * first a rotation of the camera by +90deg along the robot X axis,
             *      * then a rotation of the camera by +90deg along the robot Z axis, and
             *      * finally a translation of the camera to the side of the robot.
             * When determining whether a rotation is positive or negative, consider yourself as looking
             * down the (positive) axis of rotation from the positive towards the origin. Positive rotations
             * are then CCW, and negative rotations CW. An example: consider looking down the positive Z
             * axis towards the origin. A positive rotation about Z (ie: a rotation parallel to the the X-Y
             * plane) is then CCW, as one would normally expect from the usual classic 2D geometry.
            OpenGLMatrix robotFromCamera = OpenGLMatrix
                    .translation(mmBotWidth / 2, 0, 0)
                            AxesReference.EXTRINSIC, AxesOrder.XZY,
                            AngleUnit.DEGREES, 90, 90, 0));
            RobotLog.ii(TAG, "camera=%s", format(robotFromCamera));
             * Let the trackable listeners we care about know where the camera is. We know that each
             * listener is a {@link VuforiaTrackableDefaultListener} and can so safely cast because
             * we have not ourselves installed a listener of a different type.
            ((VuforiaTrackableDefaultListener)redTarget.getListener()).setCameraLocationOnRobot(parameters.cameraName, robotFromCamera);
            ((VuforiaTrackableDefaultListener)blueTarget.getListener()).setCameraLocationOnRobot(parameters.cameraName, robotFromCamera);
             * A brief tutorial: here's how all the math is going to work:
             * C = robotFromCamera          maps   camera coords -> robot coords
             * P = tracker.getPose()        maps   image target coords -> camera coords
             * L = redTargetLocationOnField maps   image target coords -> field coords
             * So
             * C.inverted()                 maps   robot coords -> camera coords
             * P.inverted()                 maps   camera coords -> imageTarget coords
             * Putting that all together,
             * L x P.inverted() x C.inverted() maps robot coords to field coords.
             * @see VuforiaTrackableDefaultListener#getRobotLocation()
            /** Wait for the game to begin */
            telemetry.addData(">", "Press Play to start tracking");
            /** Start tracking the data sets we care about. */
            boolean buttonPressed = false;
            while (opModeIsActive()) {
                if (gamepad1.a && !buttonPressed) {
                buttonPressed = gamepad1.a;
                for (VuforiaTrackable trackable : allTrackables) {
                     * getUpdatedRobotLocation() will return null if no new information is available since
                     * the last time that call was made, or if the trackable is not currently visible.
                     * getRobotLocation() will return null if the trackable is not currently visible.
                    telemetry.addData(trackable.getName(), ((VuforiaTrackableDefaultListener) trackable.getListener()).isVisible() ? "Visible" : "Not Visible");    //
                    OpenGLMatrix robotLocationTransform = ((VuforiaTrackableDefaultListener) trackable.getListener()).getUpdatedRobotLocation();
                    if (robotLocationTransform != null) {
                        lastLocation = robotLocationTransform;
                 * Provide feedback as to where the robot was last located (if we know).
                if (lastLocation != null) {
                    //  RobotLog.vv(TAG, "robot=%s", format(lastLocation));
                    telemetry.addData("Pos", format(lastLocation));
                } else {
                    telemetry.addData("Pos", "Unknown");
         * A simple utility that extracts positioning information from a transformation matrix
         * and formats it in a form palatable to a human being.
        String format(OpenGLMatrix transformationMatrix) {
            return transformationMatrix.formatAsTransform();
         * Sample one frame from the Vuforia stream and write it to a .PNG image file on the robot
         * controller in the /sdcard/FIRST/data directory. The images can be downloaded using Android
         * Studio's Device File Explorer, ADB, or the Media Transfer Protocol (MTP) integration into
         * Windows Explorer, among other means. The images can be useful during robot design and calibration
         * in order to get a sense of what the camera is actually seeing and so assist in camera
         * aiming and alignment.
        void captureFrameToFile() {
            vuforia.getFrameOnce(Continuation.create(ThreadPool.getDefault(), new Consumer<Frame>()
                @Override public void accept(Frame frame)
                    Bitmap bitmap = vuforia.convertFrameToBitmap(frame);
                    if (bitmap != null) {
                        File file = new File(captureDirectory, String.format(Locale.getDefault(), "VuforiaFrame-%d.png", captureCounter++));
                        try {
                            FileOutputStream outputStream = new FileOutputStream(file);
                            try {
                                bitmap.compress(Bitmap.CompressFormat.PNG, 100, outputStream);
                            } finally {
                                telemetry.log().add("captured %s", file.getName());
                        } catch (IOException e) {
                  , e, "exception in captureFrameToFile()");
    Michael P Clark
    Founding Mentor, FTC 9958
    "We're Hooked on FIRST"

  • #2
    All the classes and methods that are having errors are ones that are only in SDK v4.0. Have you been able to use other classes and methods which are only in the current SDK? Are any of the other sample programs throwing errors as well? Can you open the java files for the classes that are throwing errors to confirm that those classes exist?


    • #3
      Originally posted by Noah View Post
      All the classes and methods that are having errors are ones that are only in SDK v4.0. Have you been able to use other classes and methods which are only in the current SDK? Are any of the other sample programs throwing errors as well? Can you open the java files for the classes that are throwing errors to confirm that those classes exist?
      Thanks Noah , a little confusing as we do have other opmodes that work fine with SDK 4.0, and our DC app does show that 4.0 is installed, but certainly possible our other Opmodes are not using any methods that weren't already present in previous versions of the SDK. Even that logic is counterintuitive to us, as the Opmode still needs to import all the classes to run, even those that were present in previous SDK versions. With that, why would the Opmode correctly import some classes and not others, in this case, only those related to vuforia seem to not be importing correctly, while others are.

      We'll try to go through and see if we can narrow down the problems to those 4.0 specific classes that aren't importing, as your diagnoses certainly makes sense, still odd though that some classes load and others don't.

      Thanks for the response and good luck this season!

      Michael P Clark
      Founding Mentor, FTC 9958
      "We're Hooked on FIRST"


      • #4
        The programming mentor on my team passed this suggestion to share.

        On the 3rd photo, there is the line:
        * Instantiate the Vuforia engine */ vuforia = ClassFactory.getInstance().createVuforia(parameter s); replace it with
        this.vuforia = ClassFactory.createVuforiaLocalizer(parameters); to get rid of the compilation error. This line is from ConceptVuMarkIdentification which is last year's sample file. This getInstance() line is also in ConceptVuforiaNavRoverRuckas and by changing it with the older version of "this.vuforia", at least other teams can compile their adaptation to ConceptVuforiaNavRoverRuckas file.


        • #5
          Like Noah said, you seem to be having trouble with the 4.0 classes. Sounds like Android Studio is still using some 3.7 components

          I'm not sure what process you used to load the 4.0 SDK, but my experience with Android Studio is that it holds onto it's cached linking information WAY too hard.

          Whenever I make a major project update like from 3.7 to 4.0 I ALWAYS re-import the project once I've done all the changes.

          So, if you haven't already, I'd just import the project again. Then exit out of AS, and re-enter.

          If this doesn't help, I'd try one more thing...

          With AS closed, manually rename the main project folder (ftc-app) to something different, then open AS, and once again re-import the project.

          This folder name change forces AS to re-index, and start with all new cache information.
          As always, once you import, exit AS and restart.