-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
1/22
Surveying Graduating Seniors
and Former Graduates:
Satisfaction is only the beginning
Ronald J. Polland, Ph.D.
Assistant Director of Institutional Research, UNF
Introduction
Overview
According to by Lana Low1, Vice President of Noel-Levitz, Successfulinstitutions share three basic attributes: they focus on the needs of theirstudents; they continually improve the quality of the educational experience,
and they use student satisfaction data to shape their future directions. Studentsatisfaction studies measure how effectively campuses deliver what students
expect, need, and want. Meeting the needs of students and improving thequality of their education are priority directives at the University of North
Florida. They are included in its overall strategic plan and are consistent withits mission, goals, and objectives. In an effort to measure its effectiveness in
achieving these directives, the University requested the Office of InstructionalResearch (OIR) to gather input from its students.
In response to the Universitys request, the Office of Instructional Research(OIR) annually conducts an online survey of its graduating seniors prior to the
completion of their Spring term. Known as the Graduating Seniors Survey(GSS), the purposes of this instrument are to measure student behavior and
opinion concerning the programs and services offered at UNF, to identifyprograms and services that need to be improved, and to learn more about the
personal experiences of students during their enrollment.
Additionally, every two years, the OIR conducts a follow-up survey of UNFstudents who graduate with baccalaureate degrees one year and five years
prior to the date of the survey. Called the Former Graduates Survey (FGS), thepurposes of this survey are to learn about their experiences with the programs
and services offered at UNF, and to assess the overall relevance and impact ofthese programs and services on their current occupational and/or educational
pursuits.
1 Low, Lana. Are college students satisfied? A national analysis of changing expectations. USA Group
Foundation, New Agenda Series, February 2000.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
2/22
Since the beginning of Fall 1999, the OIR has completed two rounds of theFGS (surveying the graduating classes of 1992/1993, 1994/1995, 1997/1998,
and 1999/2000) and three rounds of the GSS (surveying the graduating classesof 2001/2002 through 2003/2004).The original administration of the FGS
marked the first time that graduates of the University of North Florida were
surveyed in a comprehensive and systematic way.
All institutions ask themselves, "Where are we now and where do we want tobe? The use of appropriate student surveys and the application of that data to
program planning and evaluation are critical to answering these questions.The following sections describe our efforts to develop graduate surveys from
which data could be used for planning and evaluation purposes.
Survey Development
The problem with Student satisfaction
One of the initial design decisions that had to be made was whether to go witha commercially available survey or to design one of our own. We decided to
develop our own for two primary reasons: (1) the commercially availablesurveys were not specifically related to our universitys goals and objectives,
and (2) we had a lot of in-house surveys upon which to base our own. At ourinstitution, college seniors are no strangers to surveys. Many of our offices
and service components like Alumni Services, Institutional Advancement, theLibrary and the Health Center were routinely surveying seniors. Additionally,
some of the individual colleges and departments were also surveying them.Besides being more relevant to our goals and objectives, the decision to create
our own survey served another purpose: by incorporating and consolidating
into our survey many of the items from other UNF surveys, the number ofdifferent surveys that a student had to answer was dramatically reduced.
In designing our survey, we reviewed as many survey items as we could fromother institutions and commercially available surveys as well as from other
UNF sources. We also studied the research literature on student satisfactionsurveys as well. We wanted to find items that did the best job of identifying
why students go to college, what they hope to achieve while in college, andwhat factors determine their choice of a specific college.
Most of the surveys we reviewed were of a Likert variety, such as those
requiring respondents to indicate their level of agreement with qualitativestatements. We decided to not use these typical Agree/Disagree scales onour surveys because of their inherent weaknesses (to be discussed below).
Instead, we chose to use scales that measured levels of quality and levels ofmagnitude or frequency.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
3/22
Another consideration we had in developing our survey was not to "reinventthe wheel." Since we already had a great deal of information gathered from
other sources pertaining to specific issues such as parking or the quality ofinstruction, we did not need to explore these issues to any great extent. For
example, after the completion of every term, all instructors (with some
exceptions) are evaluated by their students using the Instructional SatisfactionQuestionnaire (ISQ). The ISQs provide a lot of data on the perceived qualityof instruction, and consequently, we have only a few questions on our
graduate surveys about instructor quality.
From the outset, we wanted to design a survey that could provide data on how
well the university was achieving its mission, goals, and objectives as stated intheir strategic plan. Another objective was to have that survey data evaluated
and translated into an action plan.
A third objective was to design a survey comprehensive enough to cover all of
the important outcomes of the students college experience while keeping it to
a manageable length. As we learned from our initial former graduates survey,shorter is better.
The more we studied how other student satisfaction surveys were beingconstructed and used around the country, the less we found them to be
relevance to our needs and research objectives. Consequently, we decided notto model our survey after any of the student satisfaction surveys currently in
vogue. To put it another way, we were not at all satisfied with the concept ofstudent satisfaction.
According to Levitz2, the more an institution provides situations which closely
correspond to the student's expectations, the greater the level of studentsatisfaction. Additionally, the more fully an institution meets the needs of thestudents, the greater will be the level of student satisfaction. However, in
reviewing the various student satisfaction surveys in use, we found it difficultto identify the specific student expectations and needs that these surveys were
measuring. In some instances, the concept of student satisfaction was beingtreated as a concrete entity rather than as a proxy measure for describing how
effectively needs, wants, and expectations were being met.
For example, some institutions compare their student satisfaction averages to
other institutions, both individually and collectively, much in the same waythat they would compare ACT or SAT scores. Despite a number ofcommonalities among them, student needs, wants, and expectations tend to
differ even among institutions having similar traits. Because indices of studentsatisfaction are relevant only to the institution that produced them,
establishing norms like ACT or SAT scores tend to obscure their intendedusage.
2Ibid
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
4/22
The problem with scales of agreement and satisfaction
In addition to how the concept of student satisfaction was used (or abused),
we found a number of other problems in the surveys we reviewed. Inparticular, most of the problems centered on the use of Agree/Disagree and
Satisfied/Dissatisfied scales of measurement:
1. In order to fit these formats, artificial position statements are often created.
For example, one of the items we found on a survey using anAgree/Disagree scale was, "Faculty are serious about treating male and
female students equally." A more natural statement a student might say isFaculty give preferential treatment to students on the basis of gender.
2. Agree/Disagree statements force respondents to take a position For or
Against an issue which, if not relevant to them, can result in a largenumber of No opinion or Neutral responses. For example, if a small
proportion of students participate in distance learning courses, and all
students are asked if they desire more distance learning courses, the vastmajority of responses will fall in the "No response" or "No opinion" block.
3. Agreement-disagreement is not a semantic differential. For example, the
following two questions are grammatically opposite but not semanticallyopposite: There should not be designated parking areas for students, and
There should be designated parking areas for students. Students mayinterpret the first statement to mean that they should be allowed to park
anywhere they want, and interpret the second one to mean that they shouldhave their own parking areas as do faculty and staff.
4. There is no clear-cut way to interpret ones agreement or disagreementwith a statement when the proper response choices are unrelated to
agreement. For example, if a student agrees with the statement, "A varietyof intramural activities are offered," what does it mean? Does it mean that
the student is satisfied with the variety of activities offered, or is thestudent merely verifying a statement of fact?
5. Too many items using agreement scales are built upon face validity
alone. The underlying assumption in surveys of this type is that agreeingwith an item indicates a person has a specific opinion about that item.
However, the wording of a statement can restrict the range of responsessuch that agreeing (or disagreeing) with it is the only logical choice.
6. In order to classify data as "satisfaction information," many survey itemsare essentially shoehorned into a Satisfaction-Dissatisfaction format.
This is especially true for statements using the adjectives, adequate, andreasonable such as "There is an adequate selection of food available in
the cafeteria." Little thought is given to making the questions morerelevant to the intended outcomes.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
5/22
7. As an additional criticism, satisfaction surveys contain many statementsthat use the same scale format for expediency sake only, and not because
the scale fits the statement. In many instances, the scales and statementsare mismatched leading to ambiguous or even nonsensical results. For
example, using the item from #4 above: "A variety of intramural activities
are offered." This is a statement of fact that basically can be answeredwith a Yes or No response even if that variety is only three innumber and includes activities that are of no interest the student.
Answering with a Satisfaction Scale does not make sense.
8. Satisfaction surveys are good at identifying general problem areas, but do
not lend themselves to specific solutions, especially when most of theitems use the qualifiers, Adequate and Reasonable.
9. All of the statements are typically couched in positive terms thusincreasing the likelihood of a positive response bias. Not surprisingly, the
distribution of responses to satisfaction surveys is heavily skewed towards
the positive side; e.g., means averaging around five for a seven-pointscale. In a vague attempt to not be overly positive, some items includewords like Adequate and Reasonable; yet, respondents will still
interpret them as positive items. Meanwhile, other items may be off to oneside of the scale entirely, such as The instruction in my major field is
excellent. There should be a mix of positive and negative items to ensurecontent validity and internal consistency.
10.Importance and satisfaction are not comparable metrics. Importance
measures are comparative in nature while satisfaction measures are not.Rating importance is implicitly comparative; i.e., a student might think,
Parking and having a safe campus are both important to me but parking isnot as important as having a safe campus. Conversely, being satisfied
with one has little to do with being dissatisfied with the other.
11.Because they focus so heavily on student satisfaction, many surveys do
not include quantitative data such as measures of student behavior.
12.Finally, in an effort to make all items multiple-choice and easy to analyze,
free responses are too often overlooked on surveys of student satisfaction.
Satisfaction is not an outcome in and of itself, but is the consequence of anoutcome. Since satisfaction occurs as a consequence of meeting student
needs, wants, and expectations, and therefore, surveys should focus onspecific outcomes rather than on its concomitant effects. In relation to needs,
satisfaction may be a misnomer since the proper term for fully meeting a needis "satiation," not "satisfaction." In reality, meeting needs is an "all-or-
nothing" process; i.e., needs are either met or not met. We tend view theprocess of meeting needs as both quantitative and qualitative, when in fact, the
process is only quantitative. What may be either qualitative or quantitative isthe type of need met.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
6/22
The confusion occurs when we lump needs together or fail to account for allthe needs involved. For example, a steak or a hamburger can be made from
the same cut of meat. The meat in both forms can fulfill ones need for food,but there may be additional needs for which the steak fulfills but not the
hamburger. Thus, satisfaction may have more to do with the number of needs
that are met than with the degree to which any one particular need is met.
Building a better scale of measurement
Our institution has established goals and objectives for meeting the needs and
expectations of its students, but has not established any goals or objectives forincreasing "student satisfaction." The word satisfaction does not appear
anywhere in the universitys strategic plan. If the university does not considerstudent satisfaction to be an outcome, then perhaps its institutional researchers
should not either.
We approached our task of developing surveys in a way that is analogous to
needs assessment and program evaluation. We designed our survey to identifythe critical needs and expectations of students as they themselves identifiedthem. We also wanted to know if students have realistic expectations and
needs, and whether the university has realistic goals and objectives towardsmeeting them. Finally, we designed our survey to determine how well the
university is achieving its goals and objectives, and to identify the short-termand long-term impacts that the university has on its students.
In the end, our solution to the problem of defining student satisfaction as aspecific, identifiable construct was to avoid using it altogether. Instead, we
chose to use scales and response choices that were more, direct measures of
student opinion and behavior.
When constructing survey items, the best place to start is usually to includethe most basic of questions, such as those concerning age, gender, and other
demographic information. Beyond descriptive data, one of the first questionswe ask our graduates is why they went to college and why they choose our
institution. There are many different reasons why students go to college andsome may have nothing to do with getting a degree or a quality education.
Institutional researchers may think they know what students want out of theircollege experience, but unless they ask them point-blank, they can never know
for sure.
Personally, we felt that too much survey time was being devoted to measuringthe affective domain of student opinions and attitudes, and too little time on
measuring student behavior. Thus, we chose to develop our sets of questionswith particular attention to why students come to UNF, what they hope to
accomplish, what they actually do during the time they are here, and in whatways have they grown.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
7/22
When we were deciding on the type of scales to use for our surveys, we raisedthe following question, With what type of qualitative scale are students most
familiar? The answer was not immediately obvious, but it should have been:students are most familiar with scales known as letter grades. Therefore, for
some of our survey items, we chose to use scales that were comparable to
letter grades in interpretation; i.e., scales that ranged from Poor or VeryPoor to Very Good.
For some of our other qualitative items, we chose scales that more closely
matched the actual quality or quantity being measured. For example, on aseries of questions dealing with sources of academic advising, we asked
students to rate how helpful were each of the sources, using a scale rangingfrom Very Helpful to Not Helpful (a response choice of Did not use
was also available). For a series of questions asking students to rate the degreeof influence that their education had on different skill levels, we used a scale
that ranged from Very High to None. As a final example, we askedstudents to use a scale ranging from Frequently to Never to indicate how
frequently they used each of the services available at UNF.
The evaluative scales that we chose to use in our survey are both qualitative
and quantitative, and more directly relate to the underlying student needs,expectations, and behaviors that we wished to measure. Although we
constructed four and five-point scales for our surveys as most others havedone, they are neither of the Agree/Disagree variety, nor do they represent
an underlying Satisfaction/Dissatisfaction continuum. As it turned out, theonly reference to satisfaction in our surveys appears on the very last line of
our forms: it is a sentence that says, Press the SUBMIT button if you aresatisfied with your responses.
Matching survey items to goals and objectives
In developing our survey, we tried to create questions that related back to thegoals and objectives of the university. The following is an example of how
we achieved that matching. Listed below are one of the main goals and itssupporting objectives found in the university's strategic plan:
Goal: The University of North Florida will provide a campus climate that
supports student learning and student life, and one that enables faculty and
staff to fulfill their respective roles and engage in professional growth and
development
Under this goal are four objectives:
1. The University of North Florida will support student learning bymaintaining small class sizes and a high proportion of fulltime faculty
teaching courses at the undergraduate level, providing direct and personalcontact between student and faculty.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
8/22
2. The University of North Florida will demonstrate its success at improvingstudent learning by its increasing graduation rates, while taking into
account the urban nature of the institution and the proportion of studentsthat work significant numbers of hours at the same time that they attend
school.
3. The University of North Florida will improve the quality of on-campus
student life by improving support services such as orientation, retentionservices, advising, career planning, and placement; and by providing a
vital and rich intellectual life on campus through lectures, concerts,academic clubs, intramural sports, and an active intercollegiate athletic
program.
4. The University of North Florida will foster an atmosphere of civility and
respect, with faculty and staff setting the example for students by behavingin a civil, caring, and respectful manner to one another and to students.
In the first objective, there are three main focal points: reducing class size,increasing the number of full-time faculty teaching undergraduate courses,and promoting the quality of the student-teacher interaction. We included in
our survey several items that related to the first and third points (theproportion of full-time faculty was measured directly using the instructor
activity file). We had two items pertaining to class size, in addition to freeresponse items, and two items pertaining to the student teacher interaction.
Regarding the second objective, measuring increasing graduation rates wasdone directly through our student data course files and not through the survey.
To determine how many students work in addition to going to school at how
much they work, we included two questions our survey.
We included a series of questions regarding the services that the Universityprovides to students. In that section of the survey, students indicated how
frequently they made use of those services, and rated the quality of theservices. We also included questions that ask them about their participation in
extracurricular activities.
To provide data on the fourth objective, we asked students to comment on thestudent teacher relationship as well as to indicate which courses and
instructors they found to be most memorable.
We know from experience and from research that there are manycharacteristics of the instructor, the materials, the mode of instruction used,and the learning environment that go into an overall rating of quality. We had
to decide at what level of detail we needed the information to be and weighthat against how long it took the average person to complete the survey. We
decided to keep the length of the survey such that the average student couldcomplete the survey in ten minutes or less.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
9/22
Before making our survey available to students, we ran a quality and validitycheck by sending copies of the survey to key individuals in our University
community as well as having groups of students pilot test of the instrument.
Survey methods
Former Graduate Survey (FGS)
Structure of the FGS. Developed in Fall 1999, the original FGS (known thenas the Prior Graduate Survey) was divided into two sections: the first section
asked graduates to describe the nature of their current occupation and to relatetheir experiences at UNF to their current occupation. We were particularly
interested in learning how closely their current occupation was related to theirdegree major and how well UNF prepared them for their work. The second
section asked graduates to describe their experiences at UNF, with a specialemphasis towards their experiences with courses and instructors.
The original FGS (Appendix A) was a mail out survey consisting of a two-page, two-sided form printed on 8.5 x 14 paper. The use of legal size paper
and two-sided printing was necessary given the total number of items: therewere 30 numbered question statements containing a total of 137 answerable
items. Of the 137 items, 118 required closed-end responses while 19 requiredfree-responses. The survey grouped these items into seven sections: (1)
Demographics; (2) History at UNF; (3) Contributions to Personal Growth; (4)Evaluation of Courses Taken; (5) Academic Advising; (6) Quality and
Frequency of Services Used; and (7) Additional Questions.
The original FGS was mailed out to UNF students who had graduated with
baccalaureate degrees during the academic years of 1992-93 and 1997-98. TheOIR obtained mailing labels from the Office of Alumni Services containing
the names and addresses of all students who graduated with baccalaureatedegrees during those academic years.
Included with the survey was a cover letter with a detachable response card
containing our return address and a randomly generated ID code, and apostage-paid, return envelope into which respondents were to insert their
survey forms and the detachable response card. The response card was usedto keep track of who had responded from our list of graduates. The FGS was
sent to 2,436 graduates in a single mail out, and of that number, the OIR
received back 349 completed surveys for a response rate of 14.3 percent.
Data collection and analyses. Data was extracted from the mail out surveysby scanning the forms using an optical character recognition program called
Remark. This program converts scanned data into standard SPSS data files.Because the program did not have the capability to automatically transcribe
handwritten responses from different sources, the free response items had tobe coded as graphic fields and manually transcribed.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
10/22
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
11/22
The actual survey form was created using a companion program to RemarkOCR, called Remark Web survey. This program was selected because it
converted the paper and pencil templates that we used on our prior surveysinto a web-based one. The program also provides a management function and
data analysis capabilities as well. The data collected by the Web-based survey
was saved as an Excel file and imported into SPSS for further analyses.
Survey distribution. The names and e-mail addresses of all studentsgraduating in Spring were obtained from our student data course files. A bulk
mailing program called Group Mail was used to distribute the cover letter andlink to the survey. We generated random ID codes to be included in the cover
letters. The purpose of the codes was limit access to the survey, and also toprovide a way of tracking who responded to the survey. We sent a total of
three follow-up announcements.
Data collection and analyses. As mentioned above, the data was collected in a
tab-delimited text file and imported into Excel and SPSS. The analyses of the
data, the production of charts and tables, and content analyses were handled inthe same way as were the previous, mail out surveys.
Survey results
The complete results of the Fall 2002 FGS and Spring 2003 CGS appear in
Appendix D. Although the complete results for Spring 2004 are not yetavailable, here are some of the highlights of all three previous CGS surveys.
More females than males enroll and graduate at UNF each year, and,according to the last three CGS, the percentages has been increasing.
Figure 1: Distribution of Gender by Survey
107 176
81 160
101 144
0% 20% 40% 60% 80% 100%
SPRING 2004
SPRING 2003
SPRING 2002
MALE FEMALE
Our student population is growing younger as well. From last year to thisyear, there has been a major increase in the percentage of 18 to 24 year olds at
UNF.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
12/22
Figure 2: Distri ution of Age y Survey
183 68 23 8
145 65 20 11
142 72 18 12
SPRING 2004
SPRING 2003
SPRING 2002
18-24 25-34 35-44 45-54 55 and older
About 50 percent of the respondents plan on attending graduate school withinthe next year.
Figure 3: Distribution of Graduates Attending Graduate School by Survey
130 134
126 119
126 119
SPRING 2004
SPRING 2003
SPRING 2002
YES NO
More than two-thirds of all graduates receive their degrees from the Collegeof Arts & Sciences (36%) and the College of Business (32%). However, there
has been a dramatic reversal in the numbers of graduates from these collegesin the past year. In Spring 2003, 38 percent of respondents were graduates of
the College of Business while 26 percent were graduates of the College ofArts & Sciences. In Spring 2004, 49 percent of respondents were graduates of
the College of Business while 19 percent were graduates of the College of
Arts & Sciences.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
13/22
Figure 4: Distri ution of Gra uates Major Co ege y Survey
129 51 29 32 23
93 62 30 35 22
87 78 42 20 18
SPRING 2004
SPRING 2003
SPRING 2002
COAS COBA COCSE COEHS COH
Respondents were asked to indicate what type of Bachelors Degree they wereto receive. More than two-thirds (68%) indicated the Bachelor of Arts was
their first degree. About 72% also indicated the BA as their second degreechoice.
The most frequent choices of a first major were Computer and InformationSciences and Business Management.
The most frequent choice of a second major was marketing. Surprisingly,there was a shift in ones choice of major but the shift did not follow the shiftin College of graduation.
Most graduates (86%) never changed their major during their Junior or Senioryear. For the 14 percent who did change majors, all of them changed in their
senior year, and Accounting was one of their tope three most frequent choicesof a prior major.
Table 1: Distribution of Changed Majors by Survey
RANK SPRING 2004 SPRING 2003 SPRING 2002
1 BUSINESS MANAGEMENT ACCOUNTING ACCOUNTING
2 NURSING CIVIL ENGINEERING BIOLOGY
3 ACCOUNTING ELEMENTARY EDUCATION ART
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
14/22
Figure 5: Distri ution of Majors y Survey
10%
14%
9%
3%
4%
5%
8%3%
7%
11%
7%
8%
5%
7%
7%
2%
12%
2%9%
4%
5%
4%
8%
10%
8%
6%
5%
7%
16%
17%
10%
7%
ACCOUNTING
BUSINESS MANAGEMENT
CIVIL ENGINEERING
COMMUNICATIONS
COMPUTER & INFORMATIONSCI.
ELEMENTARY EDUCATION
ENGLISH
FINANCE
HEALTH SCIENCE
POLITICAL SCIENCE
PSYCHOLOGY
Big winners for 2004 were majors in Psychology, Elementary Education, andPolitical Science. There were fewer majors in Computer Science, Accounting,and Business Management.
Graduates were asked to rate the influence that UNF had on their skilldevelopment in twelve areas. Thinking logically continues to be rated the
highest and International skills the lowest. There have been gradualincreases in the rating of Understanding different philosophies and cultures,"
and Understanding written information. There was a significant decrease inthe rating for Effective computer skills.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
15/22
Figure 6: Average Ratings for Degree of Inf uence y Source an Survey
1.90
2.60
2.90
2.80
2.90
2.80
3.00
3.00
2.90
3.00
3.00
3.10
2.10
2.70
2.90
2.80
2.90
2.90
2.90
3.00
3.00
3.10
3.10
3.20
2.10
2.70
2.70
2.90
3.00
3.00
3.00
3.00
3.10
3.10
3.10
3.20
International skills
Applying scientific principles and methods
Effective computer skills
Leading and guiding others
Effective writing skills
Understanding different philosophies and
cultures
Working independently
Teamwork skills
Understanding written information
Effective communication skills
Defining and solving problems
Thinking logically
2004
2003
2002
Graduates were asked to rate courses on the basis of six characteristics: (1)
quality of their instructors, (2) size of their classes, (3) availability of requiredcourses, (4) range of courses from which to choose, (5) information about
course requirements, and (6) relevance of classes to career goals andobjectives. Size of my classes and Quality of my instructors were rated
highest across the past three years while Range of courses from which tochoose and Availability of required courses were consistently rated lowest
(although any rating of 3 or more indicates Good). The ratings wereconsistent across all categories.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
16/22
Figure 7: Average Ratings of Course C aracteristics y Survey
3.15
3.3
3.75
3.75
3.81
3.89
3.98
4.11
4.2
3.18
3.22
3.75
3.8
3.85
3.92
3.99
4.22
4.32
3.18
3.22
3.75
3.8
3.85
3.92
3.99
4.22
4.32
Range of courses from which to choose
Availability of required courses
Information about course requirements
Overall benefits vs. amount of work
required to complete
Relevance of courses to career
goals/objectives
Usefulness and value of course content
Courses were challenging and rewarding
Quality of my instructors
Size of my classes
SPRING 2002
SPRING 2003
SPRING 2004
Graduates were asked to rate the utility of seven sources of advising. Overall,
they rated "Program of study" to be the most helpful while Central advisorsin my college were rated least helpful. About 10 percent and eight percent
rated Advisers in their major and Central advisers in their college as beingNot helpful at all. Respondents indicated that student advisers were the least
used resource.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
17/22
Figure 8: Average Uti ity Ratings for Sources of A vising y Survey
2.1
2.62
2.9
2.93
3.1
2.99
3.23
2.37
2.72
2.97
2.94
3.14
3.35
3.26
2.24
2.59
2.74
2.85
3.15
3.29
3.21
Student advisors
Central advisors in
my college
Advisors in my major
Printed material,
including the catalog
Friends
Professors not
assigned as advisors
Program of study
2004
2003
2002
Respondents were asked to indicate how often they used certain types of
services at UNF as well as to rate the quality of these services. Of the serviceslisted, respondents rated the library as the most frequently used service while
services for the handicapped was rated the least frequently used.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
18/22
Figure 9: Average Usage Ratings for Service Components y Survey
0.10
0.20
0.40
0.60
0.70
0.60
0.80
0.90
0.70
1.10
1.20
1.80
2.30
2.40
0.10
0.20
0.30
0.50
0.70
0.70
0.70
0.70
0.80
1.10
1.10
1.90
2.10
2.50
0.10
0.10
0.20
0.40
0.50
0.60
0.60
0.60
0.70
1.00
1.10
1.70
2.20
2.51
Services for the handicapped
Child Development & Research Center
Safe Ride
On Campus Student Employment
Academic Resource Center
Counseling & Testing Center
Athletic events
Career Resources & Placement Service
Health Service
Recreational Services
Cultural Activities (speakers, concerts, etc.)
Academic Advising in my major
Computer Laboratories/Services
Library
2002 2003 2004
There has been a steady increase in the use of the library and a steadydecrease in the use of most of the other services.
The library and recreational services were rated the most effective or receivedthe highest quality rating while career resources & placement services
received the lowest rating. Of all the services, only the Health Serviceincreased its ratings over the past three years.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
19/22
Figure 10: Average Qua ity Ratings for Service Components y Survey
0.3
0.33
0.79
0.86
1.15
0.85
1.16
1.33
1.25
1.64
1.71
2.47
3.06
3.27
0.29
0.53
0.63
0.79
1.01
1.12
1.07
1.2
1.28
1.56
1.77
2.52
2.85
3.3
0.31
0.37
0.41
0.74
0.75
1.02
1
1.06
1.38
1.31
1.73
2.38
2.85
3.25
Services for the handicapped
Child Development & Research Center
Safe Ride
On Campus Student Employment
Academic Resource Center
Counseling & Testing Center
Athletic events
Career Resources & Placement Service
Health Service
Recreational Services
Cultural Activities (speakers, concerts, etc.)
Academic Advising in my major
Computer Laboratories/Services
Library
2002 2003 2004
Graduates were asked to indicate what changes they would recommend to
improve the quality of educational services. Most would increase the numberof courses sections available, but not the number of actual courses offered.
About two-thirds want shorter class meeting times. The most dramaticchanges from Spring 2002 to Spring 2004 are in those graduates wanting
online or distance learning classes and night or weekend classes. In Spring2002, 23 percent wanted night or weekend classes. That number dropped to 13
percent in Spring 2004. Conversely, the demand for online or distancelearning classes increased from about 8 percent in Spring 2003 to 16 percentin Spring 2004.
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
20/22
Figure 11: Frequency Distri ution of Suggeste Improvements y Survey
7%
9%
7%
11%
23%
29%
40%
35%
53%
65%
73%
1%
4%
10%
8%
15%
23%
37%
41%
52%
63%
71%
2%
8%
8%
16%
13%
27%
35%
39%
54%
63%
68%
Smaller Class Sizes
Increase # of courses
Longer Class Meeting Times
Availability Of On-Line/Distance Learning
More Weekend And/Or Night Classes
Better Scheduling Of Class Times
Range Of Available Courses
Range Of Available Majors
Reduce Frequency Of Class Meetings
Shorter Class Meeting Times
Increase # of course sections
2004
2003
2002
Asked if they could redo their college education at UNF, more than 90 percent
indicated that they would choose to go to UNF again while less than 10percent indicated that they would choose another university or college.
Figure 12: Percentage Choosing to Attend UNF Again by Survey
92.3% 7.7%
90.9% 9.1%
89.5% 10.5%
2004
2003
2002
YES NO
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
21/22
Additional Analyses
There were a wide range of responses to the four, open-ended questions at the
end of the survey. While there were as many responses as respondents toQuestion #24, Which two courses were the most valuable to you, there were
some commonalities among the other two. In response to Question #24,What educational experiences did you find most valuable? graduates
consistently indicated that their interaction with professors and teachers wasthe most rewarding. In response to Question #25 What could be improved atUNF? most graduates indicated professors, adjuncts, and teachers as areas
of concern. Most of these graduates wanted better and more consistent qualityamong instructors, better communication between themselves, their students
and their departments, and more involvement with students.
A number of graduates said that advisors and advising could be improved.
They suggested that UNF needs more qualified advisors citing lack ofknowledge, communication skills, and motivation to work with students as
deficits. Graduates also commented on the number and availability of classes.They wanted more hands-on, real-world learning experiences and more
distance learning and online courses. These free-responses are consistent withthe multiple choice items on the survey covering the same topics.
Class availability was the main issue cited in the Additional Commentssection (Question #27), followed closely by better parking. The remainder of
the comments was ones of praise for the university and its professors.
Discussion and RecommendationsThis is the third time that the online version of the survey has been conducted.
The prior response rates for the first two surveys (Spring 2002 and Spring2003) were 26 and 27.4 percent respectively. The response rate for this survey
was 30.3 percent. Although the response rate was again lower than desired,the sample was still representative of the population. Without some kind of
incentive to offer graduates for completing the survey, however, this responserate is not expected to improve very much in succeeding terms.
Overall, the results are similar across years with some notable exceptions.
There has been a shift or redistribution of degrees awarded among the fivecolleges: COBA (32% to 26%, a 6% decrease) and COCSE (17% to 12%, a5% decrease) to COEHS (8% to 14%, a 6% increase), COAS (36% to 39%, a3% increase), and COH (7% to 9%, a 2% increase).
-
7/31/2019 Surveying Graduating Seniors and Former Graduates: Satisfaction is only the beginning.
22/22
In Spring 2002, about 70% of respondents indicated that they would choosethe same major as they had done if they could do their program of studies over
again. In Spring 2003, more than 80% of respondents said that they wouldchoose the same major again. For the 27% in Spring 2002 who would change
majors, Accounting was their #1 choice. For Spring 2003, Accounting was not
mentioned as one of the choices.
In Spring 2002, most students went to college to get a better job withfurthering education ranked second. In Spring 2003, furthering education
ranked as the clear #1 choice in comparison to getting a better job, (37%versus 26%).
The number of students working and the number of work hours increasedfrom 2002 to 2003. In Spring 2002, 17 percent said that they were
unemployed and 50 percent worked 21 or more hours. In Spring 2003, 11percent said that they were unemployed and 59 percent worked 21 or more
hours.
In Spring 2002, students recommended increasing the number of coursesections and increasing availability of online courses and weekend classes as
the main ways of improving educational services. In Spring 2003, studentsalso recommended increasing the number of course sections and increasing
the availability of off-campus sites for instruction. However, they also wantedshorter meeting times and reducing the number of class meetings two
recommendations that were hardly mentioned at all in Spring 2002.
The decline in quality ratings is somewhat perplexing. It is not clear whether
this is due to a shift in the population characteristics or represents a real
perceived decline. Also difficult to explain are the increases in those choosingnot to suggest improvements which, in turn, led to the declining percentagesof each recommendation.
Although individual program areas are using the data from these surveys forprogram planning and improvement purposes, there is currently no system in
place to do this at a university level. However, that deficit will change thisSummer as the heads of the various University sectors will meet to discuss
how to utilize the data. It is hoped that a new process will be put in place torefine not only the ways that data is collected and distributed, but how it can
be used to shape policy and program development.